Hacker Newsnew | past | comments | ask | show | jobs | submit | denimboy's commentslogin


I wonder how this compares to RWKV-V5 7B

https://blog.rwkv.com/p/eagle-7b-soaring-past-transformers


I knew a dog that had a decent vocabulary. He would visibly react to words like CAR, WALK, or PARK. He also understood PARTY and MEAT and would wag his tail for PART and lick his chops to MEAT. One time we were going to a BBQ so I explained to him we were taking the CAR to a MEAT PARTY and he got super excites, wagging his tail and going to the door all while frothing at the mouth. Anecdotal but it certainly looked like he could synthesize concepts and imagine what was going to happen in the near future.

I've done a few other experiments with my foreign speaking friends and it appeared to me that dogs understand the language their owners speak primarily.


This guide explores how to evolve Emacs into a robust Python IDE. Whether you’re writing simple scripts or managing large-scale projects, you’ll find a setup that aligns with your engineering needs and reflects the power of Emacs customization.


This is the year of the Linux desktop


mergekit is the tool you need to do this

  https://github.com/cg123/mergekit
you can slice off layers and blend models with different strategies.


Mergekit is the best thing since sliced bread, as the local llm community already knows.

The dev's blog is great: https://goddard.blog/posts/

...But its not what this paper is describing. They are basically alternating models, AFAIK. Also I have other nitpicks with the paper, like using extremely old/mediocre chat models as bases:

> Pygmillion 6B, Vicuna 13B, Chai Model 6B


No because programming is more than typing and more than grinding out code. It requires understanding business and creating solutions which satisfy customer needs which requires understanding of the culture. You aren’t going to get that with some remote hands code monkey.


I think they are using the LLM as few shot learner then using that to label the rest of the training data and finally using the now fully labeled data to train a more traditional supervised classifier like DistilBert.


Sell 5500 to depress the market and then buy back at depressed price.


Right, the question will be how much are they paying these people compared to what they made previously.


I don't think anyone will come back for less than they had before. In fact, they might ask for at least 30% more.


Really, I know some sales folks who have it unemployed for six months, I think they would take a sales force job for less than they made before over being unemployed. Really, I know some sales folks who have it unemployed for six months, I think they would take a sales force job for less than they made before over being unemployed. So maybe all the laid off from salesforce found new jobs right away? They have a lot of prestige.


Oh, reality check … the average offer is down 15-20% this year. Beggars/Choosers


Keras-ocr


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: