by marcopicentini on 8/24/23, 1:26 PM with 501 comments
by daemonologist on 8/24/23, 5:34 PM
Here's some output from q4_0 quantization of CodeLlama-7b-Python (first four lines are the prompt):
# prints the first ten prime numbers
def print_primes():
i = 2
num_printed = 0 # end of prompt
while num_printed < 10:
if is_prime(i):
print(i)
num_printed += 1
i += 1
def is_prime(n):
i = 2
while i * i <= n:
if n % i == 0:
return False
i += 1
return True
def main():
print_primes()
if __name__ == '__main__':
main()
It will be interesting to see how the larger models perform, especially after community tuning and with better context/prompting.by redox99 on 8/24/23, 2:41 PM
> The Code Llama models provide stable generations with up to 100,000 tokens of context. All models are trained on sequences of 16,000 tokens and show improvements on inputs with up to 100,000 tokens.
Edit: Reading the paper, key retrieval accuracy really deteriorates after 16k tokens, so it remains to be seen how useful the 100k context is.
by up6w6 on 8/24/23, 2:29 PM
https://ai.meta.com/blog/code-llama-large-language-model-cod...
by reacharavindh on 8/24/23, 2:05 PM
I wonder if we could make such specific LLMs (one that is proficient in all things Rust, another- all things Linux, all things genomics, all things physics modeling etc) and have them talk to each other to collaboratively solve problems.
That would be a crazy future thing! Putting machines truly to work..
by Palmik on 8/24/23, 4:17 PM
by syntaxing on 8/24/23, 4:31 PM
[1] https://huggingface.co/TheBloke/CodeLlama-13B-Python-fp16
by jmorgan on 8/24/23, 5:02 PM
ollama run codellama "write a python function to add two numbers"
More models coming soon (completion, python and more parameter counts)by benvolio on 8/24/23, 2:48 PM
Not a bad context window, but makes me wonder how embedded code models would pick that context when dealing with a codebase larger than 100K tokens.
And this makes me further wonder if, when coding with such a tool (or at least a knowledge that they’re becoming more widely used and leaned on), are there some new considerations that we should be applying (or at least starting to think about) when programming? Perhaps having more or fewer comments, perhaps more terse and less readable code that would consume fewer tokens, perhaps different file structures, or even more deliberate naming conventions (like Hungarian notation but for code models) to facilitate searching or token pattern matching of some kind. Ultimately, in what ways could (or should) we adapt to make the most of these tools?
by lordnacho on 8/24/23, 3:54 PM
Is anyone working on a code AI that can suggest refactorings?
"You should pull these lines into a function, it's repetitive"
"You should change this structure so it is easier to use"
Etc
by Draiken on 8/24/23, 4:22 PM
I absolutely love the idea of using one of these models without having to upload my source code to a tech giant.
by scriptsmith on 8/24/23, 3:05 PM
I see the GitHub copilot extensions gets a new release one every few days, so is it just that the way they're integrated is more complicated so not worth the effort?
by mymac on 8/24/23, 3:48 PM
by modeless on 8/24/23, 2:24 PM
by ilaksh on 8/24/23, 2:33 PM
I would argue that many teams will have to reevaluate their LLM strategy _again_ for the second time in a week.
by WhitneyLand on 8/24/23, 9:01 PM
I guess since Xcode doesn’t have a good plug-in architecture for this I began experimenting more with a chat interface.
So far gpt-4 has seemed quite useful for generating code, reviewing code for certain problems, etc.
by 1024core on 8/24/23, 8:01 PM
by gorbypark on 8/24/23, 7:28 PM
It might be the quantization or my lacklustre prompting skills affecting it, though. To be fair I did get it to output a little bit of useful code after trying a few times.
by TheRealClay on 8/25/23, 1:27 AM
by KaiserPro on 8/24/23, 9:11 PM
Where its a bit shit is when its used to provide auto suggest. It hallucinates plausible sounding functions/names, which for me personally are hard to stop if they are wrong (I suspect that's a function of the plugin)
by natch on 8/24/23, 3:59 PM
by jasfi on 8/24/23, 2:56 PM
by ilaksh on 8/24/23, 2:43 PM
by andrewjl on 8/24/23, 3:54 PM
I haven't yet read the whole paper (nor have I looked at the benchmark docs which might very well cover this) but curious how these are designed to avoid issues with overfitting. My thinking here is that canned algorithm type problems common in software engineering interviews are probably over represented in the training data used for these models. Which might point to artificially better performance by LLMs versus their performance on more domain-specific type tasks they might be used for in day-to-day work.
[1] https://github.com/openai/human-eval
[2] https://github.com/google-research/google-research/tree/mast...
by msoad on 8/24/23, 2:09 PM
by dangerwill on 8/24/23, 3:35 PM
You can scream that this is progress all you want, and I'll grant you that these tools will greatly speed up the generation of code. But more code won't make any of these businesses provide better services to people, lower their prices, or pay workers more. They are just a means to keep money from flowing out of the hands of the C-Suite and investor classes.
If software engineering becomes a solved problem then fine, we probably shouldn't continue to get paid huge salaries to write it anymore, but please stop acting like this is a better future for any of us normal folks.
by MuffinFlavored on 8/24/23, 2:35 PM
by e12e on 8/24/23, 5:00 PM
I see both vscode and netbeans have a concept of "inference URL" - are there any efforts like language server (lsp) - but for inference?
by pmarreck on 8/24/23, 9:16 PM
by robertnishihara on 8/26/23, 12:58 AM
by brucethemoose2 on 8/24/23, 2:30 PM
https://ai.meta.com/research/publications/code-llama-open-fo...
by naillo on 8/24/23, 2:46 PM
by braindead_in on 8/24/23, 2:56 PM
by awwaiid on 8/25/23, 4:35 AM
by pelorat on 8/26/23, 11:17 AM
by bick_nyers on 8/24/23, 3:26 PM
by dchuk on 8/25/23, 2:37 AM
by 1024core on 8/24/23, 3:26 PM
What?!? No Befunge[0], Brainfuck or Perl?!?
[0] https://en.wikipedia.org/wiki/Befunge
/just kidding, of course!
by jtwaleson on 8/24/23, 8:32 PM
by akulbe on 8/25/23, 12:27 AM
I got a bridge, but it was the wrong size.
Thanks, in advance.
by dontupvoteme on 8/24/23, 9:34 PM
by gdcbe on 8/24/23, 4:40 PM
by ai_g0rl on 8/25/23, 1:05 AM
by RobKohr on 9/1/23, 9:56 PM
by rafaelero on 8/24/23, 7:31 PM
by mdaniel on 8/24/23, 3:40 PM
by WaitWaitWha on 8/24/23, 5:02 PM
Asking for purposes of educating non-technologists.
by eurekin on 8/24/23, 2:34 PM
by m00nsome on 8/25/23, 7:22 AM
by KingOfCoders on 8/25/23, 8:10 AM
by born-jre on 8/24/23, 4:49 PM
I can see some people fine-tuning it again for general propose instruct.
by bryanlyon on 8/24/23, 4:27 PM
I do wonder about how much use it'll get, seeing as running a heavy language model on local hardware is kinda unlikely for most developers. Not everyone is runnning a system powerful enough to equip big AIs like this. I also doubt that companies are going to set up large AIs for their devs. It's just a weird positioning.
by bracketslash on 8/24/23, 6:01 PM
by the-alchemist on 8/24/23, 3:21 PM
by maccam912 on 8/24/23, 2:23 PM
by marcopicentini on 8/24/23, 4:00 PM
by binary132 on 8/24/23, 3:04 PM
by jerrygoyal on 8/25/23, 5:46 AM
by waitingkuo on 8/24/23, 1:35 PM
by mercurialsolo on 8/24/23, 4:34 PM
by Dowwie on 8/24/23, 3:14 PM
by gw67 on 8/24/23, 1:33 PM
by praveenhm on 8/24/23, 5:54 PM
by nothrowaways on 8/25/23, 1:17 AM
by likenesstheft on 8/24/23, 3:07 PM
by jrh3 on 8/24/23, 9:48 PM
by Someone1234 on 8/24/23, 8:01 PM
- Easy plug & play model installation, and trivial to change which model once installed.
- Runs a local web server, so I can interact with it via any browser
- Ability to feed a model a document or multiple documents and be able to ask questions about them (or build a database of some kind?).
- Absolute privacy guarantees. Nothing goes off-machine from my prompt/responses (USP over existing cloud/online ones). Routine license/update checks are fine though.
I'm not trying to throw shade at the existing ways to running LLMs locally, just saying there may be room for an OPTIONAL commercial piece of software in this space. Most of them are designed for academics to do academic things. I am talking about a turn-key piece of software for everyone else that can give you an "almost" ChatGPT or "almost" CoPilot-like experience for a one time fee that you can feed sensitive private information to.
by lolinder on 8/24/23, 2:23 PM
The only thing I've been able to think is they're trying to commoditize this new category before Microsoft and Google can lock it in, but where to from there? Is it just to block the others from a new revenue source, or do they have a longer game they're playing?
by rvnx on 8/24/23, 3:27 PM
In the meantime, we are still waiting for Google to show what they have (according to their research papers, they are beating others).
> User: Write a loop in Python that displays the top 10 prime numbers.
> Bard: Sorry I am just an AI, I can't help you with coding.
> User: How to ask confirmation before deleting a file ?
> Bard: To ask confirmation before deleting a file, just add -f to the rm command.
(real cases)
by 6stringmerc on 8/24/23, 2:47 PM
No thanks, going back to Winamp.