Well, it depends on the audience of your software. Bloomberg Terminal is awesome, but has a steep learning curve. Most people who learn to use it are learning it with the specific intention of finding a job where they are paid to know how to use the Bloomberg Terminal. I agree, the UI is awesome for that, but if your startup is an app to find a dog walker or something, the vast majority of your potential users are going to be turned off by the Bloomberg Terminal interface.
Yeah but even if your app is the new Uber for dogs the goal should still be to let users accomplish their task as fast as possible and not some app engagement metric.
What LLM are you currently using? And how do you prevent against hallucinations/drift when generating content of this length? Do you start by asking the model to generate an outline of a book, and then have it expand on each chapter with more detail? Awesome project
Currently using GTP-4o combined with Perplexity API context infusion for real-time knowledge (this also reduces hallucination for the most part). And yes, starting with outline and then writing chapter by chapter while doing real time research. After initial completion there’s another editing round to make everything more coherent.
I still use a 2014 macbook air as my home server. It was my daily driver for 7 years. No complaints, still works perfectly. Haven't been careful with it either.
They're grown from stem cells. Neuro-ethics (especially the ethics around using brain organoids) is an entire field of research. Although brain organoids exhibit neural connections and electrical activity, they have so far failed to form even basic synaptic circuits — without which consciousness is probably impossible. But of course we should take every precaution to ensure that they DON'T form consciousness. But that's still a ways away.
What's the harm if they do form consciousness? Is the concern they will develop a willpower and work towards freeing themselves, or is it just a "don't enslave sentient beings" thing?
Are there any good resources for understanding models like this? Specifically a "protein language model". I have a basic grasp on how LLMs tokenize and encode natural language, but what does a protein language actually look like? An LLM can produce results that look correct but are actually incorrect, how are proteins produced by this model validated? Are the outputs run through some other software to determine whether the proteins are valid?
Proteins are linear molecules consisting of sequences of (mostly) 20 amino acids. You can see the list of amino acids here: https://en.wikipedia.org/wiki/Amino_acid#Table_of_standard_a.... There is a standard encoding of amino acids using single letters, A for alanine, etc. Earlier versions of ESM (I haven't read the ESM3 paper yet) uses one token per amino acid, plus a few control tokens (beginning of sequence, end of sequence, class token, mask, etc.) Earlier versions of ESM were BERT-style models focused on understanding, not GPT-style generative models.
Agreed, would be interested if someone with more knowledge could comment.
My layman's understanding of LLMs is that they are essentially "fancy autocomplete". That is, you take a whole corpus of text, then train the model to determine the statistical relationships between those words (more accurately, tokens), so that given a list of tokens of length N, the LLM will find the next most likely token for N + 1, and then to generate whole sentences/paragraphs, you just recursively repeat this process.
I certainly understand encoding proteins as just a linear sequence of tokens representing their amino acids, but how does that then map to a human-language description of the function of those proteins?
Most protein language models are not able to understand human-language descriptions of proteins. Mostly they just predict the next amino acid in a sequence and sometimes they can understand certain structured metadata tags.
Can they understand the functional impact of different protein chains, or are they just predicting what amino acid would come next based on the training set with no concern for how the protein would function?
The way you would use a protein language model is different from how you would use a regular LLM like chatgpt. Normally, you aren't looking for one correct answer to your query but rather you would like thousands of ideas to try out in the lab. Biologists have techniques for trying out thousands or tens of thousands of proteins in a lab and filtering it down to a single candidate thats the best solution to whatever they are trying to achieve.
Get a sleep study done. Lofta will ship you an at-home test. Sleep apnea will leave you feeling unrefreshed no matter how much sleep you get, and it will only get worse without a CPAP or other aid.
I would love to see a comparison of the average amount of trash (especially plastic waste) the average American homeowner produces vs the average European homeowner. I bet the difference is astonishing.
But probably everything-consumption is really high, with a toxic transportation system (sand for roads), land use, electronics use (rare earths), etc. Fucking up the planet is the easiest way to increase GDP.
$60 + lost income for a student who works outside school hours and also has to take the test outside school hours + transportation which could be over an hour in rural areas and you can see how this could quickly add up, especially for a teenager who may be smart but may not see the proper cost/benefit of skipping the SAT
All this seems like a small investment compared to properly preparing for the test. An average student would probably want to study at least 10 hours if they care about their test score.
> especially for a teenager who may be smart but may not see the proper cost/benefit of skipping the SAT
The SAT is for students who want to attend college. If $60 and a two hour round trip is going to tilt the scale on whether you spend 4 years in college, then from the college's point of view, you may not be the type of applicant they are looking for.
You are looking for corner cases within corner cases at this point. I know it's uncool to assign the responsibility of children to their parents these days... but if you subscribe to the idea that every system in place has to cater to every possible living situation, you'll just end up with one that doesn't work well for anyone, as evidenced by this post.
Absolutely agree. If these tests are not offered in the school during normal school hours, students should receive compensation for taking the test. It's bonkers to me that students have to pay to take it. Travel should also be provided if the test cannot be offered in the student's school. I think that would be a much fairer way to close the equity gap for standardized testing, rather than remove one of the few ways a student from a low-income background can prove their academic worth.
Not a dumb question, I’m assuming it’s been treated with flame-retardant chemicals, but I’m surprised that it’s fireproof-ness wasn’t addressed in the article.
Since they are the tallest thing around, wind turbine blades get struck by lightning regularly.
Fiberglas is an insulator. Aluminum is a good conductor. But carbon fiber is a resistor. Carbon fiber blades have to have embedded conductors to prevent from being blown up by lightning strikes. So there's a layer of aluminum foil or similar to provide a low-resistance discharge path.
Same problem as carbon-fiber aircraft, with similar solutions.