Hacker Newsnew | past | comments | ask | show | jobs | submit | more sp332's commentslogin

I was browsing test kits where you send a sample back to a lab. They're $79 - $300 and advertise 1 or 2 parts per trillion. Looks like getting to 0.02 ppt requires some very specialized equipment, and would probably be optimized for continuous monitoring of a water supply.


Fine tuning can be useful if you need to generate lots of output in a particular format. You can fine-tune on formatted messages, and then the model will generate that automatically. That could save a bunch of tokens explaining the output format in every prompt.


You can use structured generation instead of fiddling with the prompt, which is unreliable. https://github.com/outlines-dev/outlines


Does this Python package control the LLMs using something other than text? Or is the end result still that that Python package wraps your prompt with additional text containing additional instructions that become part of the prompt itself?


Looks like it actually changes how you do token generation to conform to a given context-free grammar. It's a way to structure how you sample from the model rather than a tweak to the prompt, so it's more efficient and guarantees that the output matches the formal grammar.

There's a reference to the paper that describes the method at the bottom of the README: https://arxiv.org/pdf/2307.09702


The output of the LLM is not just one token, but a statistical distribution across all possible output tokens. The tool you use to generate output will sample from this distribution with various techniques, and you can put constraints on it like not being too repetitive. Some of them support getting very specific about the allowed output format, e.g. https://github.com/ggerganov/llama.cpp/blob/master/grammars/... So even if the LLM says that an invalid token is the most likely next token, the tool will never select it for output. It will only sample from valid tokens.


No it limits what tokens the LLM can output. The output is guaranteed to follow the schema.


How much money do you have?


What does it cost me to p .95 reliably differentiate 0.04 ppt from 0.02 ppt?


In water? Doesn't seem like an issue.


In pure water it's not an issue. In drinking water which has all sorts of "stuff", getting very to that level of precision isn't easy.


The doctor in the documentary tells him that his liver is damaged like an alcoholic's. He just let people assume it was from the food.


He said it was this specific performance that convinced Weird Al that he was the right guy. Also the movie is great.


Since the 777 is older, and this has never happened, and they are giving airlines 5 years to make the change, it doesn't seem likely to be an actual problem.


Let's hope so, it's not a problem you can afford to be around even once.


It's not "reportedly".


so it wasn't reported?


Not by any serious outlet. It's a Twitter rumor started by a nobody... Another reason why letting people pay for blue checks was an incomprehensibly stupid idea.


As the body of the article says repeatedly, it's only a rumor.


It’s rumoredly.


I don't remember Currents or Spaces.


Similar to the deal they got from Microsoft, >$10B on paper but a lot of that is in the form of Azure credits.



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: