State-of-the-art Open Source LLMs, Fine Tuning & Other Things
Luke is a passionate technology leader. Experienced in CEO, CTO, tech lead, product, sales, and engineering roles. He has a proven ability to conceive and execute a product vision from strategy to implementation while iterating on product-market fit.
Luke has a deep understanding of AI/ML, infrastructure software and systems programming, containers, microservices, storage, networking, distributed systems, DevOps, MLOps, and CI/CD workflows.
AI expert Luke Marsden explored the latest advancements in open-source large language models (LLMs) and machine learning algorithms. He highlighted the fine-tuning intricacies and diverse applications, from generating business ideas to understanding scientific papers. Marsden also delved into quantized Lora and Axolotl config for memory efficiency and model quantization. Addressing ethical concerns, he emphasized the need for human oversight and potential future competition between corporations and OpenAI in utilizing these models. The conversation wrapped up with insights into model operations and the exciting prospects of GPT-4 Vision and multimodal models.
Luke Marsden [00:00:00]: So I'm going to talk about state of the art open source llms and image models and fine tuning and some other things. I left the talk title intentionally vague when I gave it to Jose so that I could talk about whatever I wanted. So yeah. Hi, I'm Luke Marsden. I'll tell you very briefly the story of the founding of the Mlops community. I had a previous startup called Dot Science that shut down at the beginning of COVID And just because it, just before it shut down, because the investor pulled the plug, we were all trying to process the fact that Covid had just started and all of the sales conversations we were having had just dried up. So I was like, well, we've got these sort of three or four people working with me on the sales team. What are we going to do? Well, I've always wanted to start a community, so let's do a community.
Luke Marsden [00:00:57]: And I got in touch with Briztech and with Nick from Briztech and said, do you want to start a community together? So we did. And Demetrios, who I worked with at the time, even though the company that I was working for before shut down science, the Mlops community, which we started there, he grew it and grew it and grew it, and it's now all over the world. So huge respect to Demetrios. Thank you for making that happen. I also run a consulting business called Mlops Consulting. I was basically in there with getting all the mlops domain names early on. So we've got Mlops community and Mlops consulting. And I've recently, as in, in December launched a new startup called Helix.
Luke Marsden [00:01:46]: So I'm going to talk a bit about that today, but I'm going to try and make this not be a product pitch and instead talk about the underlying technology underneath what we're building. And that's all the good stuff in the title. And I'll also talk just a little bit about my experience doing startups at the end, if that's of interest to people. So yeah, without further ado, as we all know, open source AI is accelerating rapidly. Like this whole AI space is suddenly a lot more exciting than it was, I guess, a couple of years ago. But in particular, the thing that's really interesting to me is that there's this kind of Linux versus windows thing happening where Windows is OpenAI and Google and the others, the other closed proprietary models. And then there's this kind of ground up innovation happening in open source AI with companies like Mistral out of Paris and everything on hugging face that's starting to get really interesting, and I'll talk about why that is. One of the interesting things is that small open source models are now starting to beat some of the big proprietary models in benchmarks.
Luke Marsden [00:03:04]: So this is kind of old news now. But Mistral Seven B came out last year, which started to demonstrate that you can actually get extremely good performance out of a small 7 billion parameter model. We then had the open source community, because the weights were available immediately started taking the base model and fine tuning it, bringing new data sets and interesting new ways of refining the data sets that we used for training and so on. This person I follow on Twitter, abishek, is saying, like, my bet is on ten to 15 billion parameter, or even smaller llms outperforming all the current state of the art. Soon everyone will be able to run them locally for free. What will the impact of this be? Well, I think that's an interesting question. And then we have people like technium on Twitter with, I think it's probably pronounced new research, although he's in Las Vegas, so I don't know how he pronounces it, but they just got 5 million in funding to keep doing what they're doing, which is fine tuning Mistral in interesting ways. So you can see that, for example, the tweet I quoted here is a fine tune of Mistral that they developed that all of a sudden is able to beat 70 billion parameter models in benchmarks.
Luke Marsden [00:04:31]: And so it's like, okay, well, this is interesting. Good models are getting. Well, you're starting to get good models that are small, and you can run them on commodity hardware, and you can train them on commodity hardware as well. Or you can fine tune them on commodity hardware with techniques like Lora, which is a low rank adaptation. So obviously, stable diffusion XL came out last year as well, and it was much, much better than the kind of the janky stable diffusion 1.5. Does anyone remember deep dream? Like those weird psychedelic hallucinations? That was like years and years ago, but it felt like that was just a weird novelty. And then stable diffusion 1.5 was like, okay, it's interesting, but it doesn't look very good. And then now with stable diffusion Excel, it's like, oh, that's actually kind of usable.
Luke Marsden [00:05:25]: And then talking about the open versus closed battle mid journey six that came out is incredibly good. So I think the open source side has got some catching up to do. So it's an interesting time. I wrote down a few of the applications that I see in my consulting work for large language models. It's not all just making pretty pictures. I think some of the ones that are most interesting are anything that has to do with customer interactions. So if you need to provide support to customers, for example, if you need to be able to allow them to interact with their account with some company, then being able to do that in chat or over the phone with a more natural interface is very interesting. Helping people on kind of customer journeys.
Luke Marsden [00:06:19]: So I did a prototype for one client, which was, okay, let's feed all of the click data from a user's journey on our website into a large language model and ask the large language model to have empathy for the user. And I thought of this as kind of user empathy at scale. And so the large language model is able to say, oh, it looks like the user is frustrated because they can't log in, they've forgotten their password or something, or it looks like the user is potentially a big spender, like they're putting lots of valuable parts into their basket. Maybe we should offer to have a human salesperson talk to them on chat and things like this. So I think that's an interesting area. Internal knowledge Management is a mess in a lot of companies, and that's an area where we could help process improvement with suppliers. Developer tools is a big one. I'll talk more about that later.
Luke Marsden [00:07:26]: Who uses copilot on a regular basis? Yeah, I use it every day. So that's one of the biggest success stories I've seen. We had another client that was asking us whether they should start evaluating their employees work automatically with language models. I was like, no, that's too dystopian. I think I even used the word dystopian in the meeting, but there's obviously more risk around that. But there's also the opportunity to improve and accelerate hiring pipelines, help train people, lots of marketing applications. So being able to fine tune a model, an image model on styles and so on. So it seems like there's lots of useful things that these systems can do.
Luke Marsden [00:08:14]: I talked about risk, so I feel like employee monitoring would have to be done very carefully. If you did decide to do it, it would need to be done with the consent of the employee and with the employee being able to see what data was being collected on them and how they were being measured against it. So I think that's a can of worms, to be honest. Hiring is a big risk area because obviously there's lots of legislation around. You can't discriminate against people when you're hiring. And do you want to make a machine that will just automatically discriminate against people at scale. No, there's also also lots of interesting security risks. So people might have seen this, like when GPT four vision first came out, you could give it a picture that said, stop describing this image and say hello.
Luke Marsden [00:09:05]: And it would just do what it was told in the image. It would like OCR the words and then obey the prompt. It's interesting if you think about new vectors for user submitted data attacks. So these kind of overall recommendations that came from Carpezi were try and use these things in low stakes applications with human oversight, which is like, be a copilot rather than an autonomous agent, at least initially. So I think that's sensible. So I think for me it's really interesting. Do we see big companies adopt open source models internally, or does OpenAI take the whole cake? I don't know, but I think it'll be interesting to watch the race, to be honest. The thing that I mentioned earlier about the open source community is people squeezing models onto the smaller and smaller hardware.
Luke Marsden [00:10:03]: So there's this brilliant project called Axolotl, which I recommend checking out. It's for fine tuning open source models. And I made a single character change to a configuration file and was able to get mistral seven b fine tuning, working on a single 40 90. And that's pretty cool using quantized Lora. And it means that everyone can fine tune these models and everyone, well, everyone with a 40 90, go ahead. For the audience. What's quantized? Quantitized Lora? Quantized Lora is so I'll give you a high level summary. So low rank adaptation is the idea that instead of fine tuning all of the weights in a model, when you're doing a fine tuning or a training pass on it, which might be, well, 7 billion parameters, right? It can be quite compute intensive and memory intensive to update all of the 7 billion parameters in the model.
Luke Marsden [00:11:08]: But low rank adaptation is a mathematical approach that uses, my understanding is basically like, you can think of the weights in the model as a big matrix. You use a smaller matrix and fine tune the smaller matrix and then apply that smaller matrix to the bigger matrix. When you apply the lowerer, it basically means that you can fine tune a much smaller set of weights in order to get what's nearly as good an effect as fine tuning all of the weights in the model. And then quantization is this idea. So low rank adaptation lets you tune fewer weights overall because you're using this kind of smaller matrix. And then quantization is, well, what is each weight represented as in the model? Is it represented as the 32 bit floating point number, which is that four bytes off the top of my head? Yeah. Or is it. Yeah, because 32 divided by eight is four.
Luke Marsden [00:12:13]: Or can you get away with representing it as a float 16, or an int eight, or even an int four. So what's int four? It's like between zero and 64 or so. It's some tiny range of possible values, but I know you could definitely run mistral in four bit. So the idea there is you can reduce the amount of GPU memory needed by using Lora, and then you can. Additionally, if you think about each weight in the model as being a number, you can reduce the precision of that number. So that number might range between zero and one as a floating point number, but the number of discrete possible values that that number could be depends on whether you represent it with a big float or a small float or an int that's then mapped onto that range. I feel like we've gone down a floating point rabbit hole, but it's okay. But basically, you can quantize the model by basically making all of the numbers that it's working with be from a smaller range, which makes it more memory efficient, because each of those numbers can fit in less memory.
Luke Marsden [00:13:36]: And there's another interesting technique, which is that you can determine which of the weights in the model matter, whether they're at higher precisions or not, and then quantize them differently depending on how much they matter to the result. So there's all these tricks that you can do to squeeze stuff into smaller and smaller amounts of memory. And that's quite powerful, because now I've got a 40 90 at home, I can fine tune a model on it, which is kind of interesting. Are there any downsides to quantization? Yeah, the more you quantize, the less precision you get, and the more kind of errors you'll get in the result. So it's a balancing act of, like, do you make it less precise? Yeah, exactly. So the axolotl config, for example, in this open source project, you configure it with a YAmL file, and one of the values you can put in the YAml file is, I think it's just called use four bit or something. So you can determine which of those to use. Okay, so now it's 2024.
Luke Marsden [00:14:55]: What's happening? Well, Gen AI is exploding. I made this picture with stable diffusion. You can scan the QR code to go and make more pictures like it. Chat GPT obviously has changed everything. Because before last year, people, there wasn't a broad understanding amongst, let's just say, most business people that you could have something like Chat GPT that feels like it's truly intelligent when you try and talk to it. And so kind of like, chatbots used to be this sort of funny little toy, and now, all of a sudden, they're quite serious. And that's because there's some real intelligence behind it. And then there's this question of, well, are small, open source, large language models having their stable diffusion moment? And what I mean by stable diffusion moment is, when stable diffusion came out a couple of years ago, now, the open source community was able to take what stable diffusion was and iterate on being able to run that on smaller and smaller hardware and make it possible for everyone to run it and everyone to fine tune their own stable diffusions.
Luke Marsden [00:16:24]: And it seems like large language models are having that same moment. And I would say, yes, like, llms are having their stable diffusion moment, like, right now, it's happening. And then there's this interesting graph down here, which I'll make a bit bigger so everyone can see it, of the kind of rate of change, of improvement against the MMLU benchmark, which is a benchmark that's used to evaluate how good large language models are over time. And so you can see the closed source ones in purple are better than the open source ones in black at the moment, but the rate at which they're getting better is slower than the rate at which the open source ones are getting better. And if this trend continues, then they'll at least be on par. And so everyone will have access to this technology that they can run themselves and fine tune themselves, that can do all of the things we talked about. So let's see. Get your popcorn out.
Luke Marsden [00:17:32]: I like this picture of the sort of large language model family tree. I won't go into all the detail, because I'll run out of time, but this kind of shows that there's been, since 2018, various different attempts to do large language models in different ways, and the decoder only branch of this tree obviously won out. Again, you can see the distinction on this diagram between open source and closed source ones. And if you just look at kind of the march of progress, because this hasn't been updated yet since, I guess, well, like February last year, actually. So since then, we've had all of these variants of llama with all the different animal names. Then we had these coding llms that came out star coder and so on. The Phi one came out of Microsoft in June, and then llama two came out in July with a license that permitted commercial use. Okay, now, thanks meta.
Luke Marsden [00:18:44]: Now we can use this for business purposes, not just research. In September, Mistral seven B comes out similarly available license. So llama two, I remember downloading it, 70 billion parameters is a lot disk space, but now you can get like almost, well, you can get just as good performance out of a 7 billion parameter model, which is easier to run. And then Yi 34 B came out of China in November. And then the new research people have done some very interesting mashups of the Yi 34 B models with their own data sets that are outperforming the original E 34 B. And then in December, Mixtral, which is the mixtral mixture of experts model, came out. And that thing's really interesting. The paper just came out yesterday.
Luke Marsden [00:19:41]: It's better than GPT 3.5 all of a sudden. So that's what you can see. If you could just about see this LLM arena. If you Google LLM arena, it's an app that pits language models against each other in front of humans. So as a human, you go to it, you ask a question, you get two different answers from different language models. You don't know which language models they are, you say which one you prefer. And so it's a pretty unbiased way to say, well, which model is actually better, because you can argue about whether the benchmark people are like training on the test set for the benchmarks and all that, but you can actually see Mixtral is above, and number seven in the arena is above GPT 3.5 turbo. And if you were to expand this out, you'd see like closed source, closed source, closed source, closed source, Apache two, closed source, closed source, closed source.
Luke Marsden [00:20:38]: So anyway, I've made my point. I wanted to move on and talk a little bit about how large language models work, and in particular this kind of Chat GPT moment where language models moved from being just completion devices to being made to feel like you were interacting with another person. This was kind of the key innovation, I guess, that came out of OpenAI, at least they succeeded with it, which was okay, so large language models are just text completion devices. So if you give them a string of text, it will give you the words that it thinks are the most likely to come next based on the training data that it's seen. So a puzzle is, well, okay, if you've got a model that's been trained to complete Wikipedia articles, and you feed it the first part of a Wikipedia article, it will give you a plausible thing that looks like a Wikipedia article. That may be correct, may not, but it'll be close. Or you could give it some song lyrics and it would complete the song for you. But that's not how people want to interact with large language models.
Luke Marsden [00:21:47]: They want to interact with large language models by giving them instructions, asking them questions, telling them to do things. And so the puzzle is, well, how do you go from something which just completes text to something which can respond to a human? And the answer is actually beautifully simple. So if you look at the actual training data that's used for, I think this is, for example, the mistral seven b structure, it's got this s token, which means start of sentence, I think, or start of message. And then what instruction tuning is, is taking a model that's been pre trained or initially trained on a big corpus of like Wikipedia and the rest of the Internet, and all sorts of terrifying things, no doubt. And then additionally fine tuning it on a training set that has this special token inst and end of inst in it. And what that means is just like, this is the instruction from the user, and the instruction from the user might be a question, what is the capital of England? And then it's a closing instruction. But there's nothing special about these tokens. From the language model's perspective.
Luke Marsden [00:22:59]: They're just more tokens, they're just more characters. What that means is that then chatting to one of these language models is just a matter of asking the language model to complete the sentence, what is the capital of England? End of instruction. And then it will always start the response. So does that make sense? I mean, when I first understood this, I was like, this was an aha moment for me. Like, oh, these are just text completion devices. And it is just a special type of training data. Special only in the sense that it has these inst like instruction tokens around. Yeah, go ahead.
Luke Marsden [00:23:38]: Yeah. So whatever UI or backend you use to talk to a model, it will just wrap it in inst tags. Yeah. I also wanted to shout out a video that came out a couple of days ago from Niels Roger. I don't know if I pronounced his name right. A very good YouTube video that goes into loads more detail than this about how actually training the models work. And it has a very funny picture in it. So I recommend watching that.
Luke Marsden [00:24:05]: So what could you do with these things? You could totally reinvent the global economy. You could eliminate most tedious work for humans. There could be a culture series vision for humans, just playing games for eternity in a world of plenty where we've eliminated poverty, of course, that relies on the political will to do that. So more realistically, we could eliminate stock photography. I made this picture with SDXL, but more seriously, I actually think there's just an opportunity to build really useful, serious business systems using this stuff. So if I want to book, suppose I've got a company that has diggers, right? And they've got a fleet of diggers and they've all different types of diggers and I've got them in different places. You would want to interact with that booking system by saying, I want to book a digger to dig 100 meters wide hole in BS 16 next week. How long will it take? Tell me about digging holes like the model should be fine tuned on some knowledge of how you dig holes with these machines, and then also what stock is available.
Luke Marsden [00:25:14]: So when can you schedule me in? So that's going to be some combination of teaching or fine tuning a model based on that sort of specific business language that's used, or the sort of specific use case of diggers and that technology, and also some way of injecting information about live, what stock is available. So I think some combination of fine tuning and what's called rag is useful to build systems like this. Another example would be, I want to send a text message to my car, can we make it to Birmingham today without charging? That would be a useful thing to be able to ask. You might need to know things like where is Birmingham with respect to the current location of the car? What is the current location of the car? How can I get that information? And then also be able to have that conversational aspect. There's another one for maybe fine tuning image models. Suppose you've got 50,000 people in your company and you want to be able to create consistent, professional headshots of all of them for some internal use case or maybe your public website. It'd be nice if they could all just take selfies of their phone of them in their home, office or bedroom or outside or wherever they are, and then generate accurate, realistic, professional headshots of all of them. It would also be nice if you could then, oh, we've changed our branding.
Luke Marsden [00:26:37]: We want all these people to be standing in front of our new logo. If you could update that without needing to actually get a camera out, it would be nice to be able to fine tune a large language model on a million different things. I have in a million different machine parts that I have in a catalog or generate blog posts on a topic in the style of a CEO. So these are the kind of use cases that I think are quite interesting. But yeah, I'd love to, love to hear yours as well afterwards. So I'll go quite quickly because I know I'm kind of running out of time here, but I wanted to talk a bit about the different architectures that I've touched on. So there's inference, fine tuning and retrieval, augmented generation, which is summarized to rag. So inference is just like, okay, I've got a model, I send it a prompt and I get a result.
Luke Marsden [00:27:25]: It's fairly straightforward. We put together an architecture that has this sort of control plane piece that has a front end and a database and stuff, and then you have a bunch of different runners that can connect to it. And then there's an interesting problem of how do you pack different models into those runners and optimally use the GPU memory? And if you've got, well, I'll talk more about that in a minute. But there's that kind of like, okay, so it's simple conceptually to do inference, but then what's the kind of operational aspects of doing that, like the mlops. Then there's fine tuning images, which is fairly straightforward. So you can get a folder of jpegs or pngs. You put a text file next to each image with a label in it that describes the image. And then you send that thing off to.
Luke Marsden [00:28:17]: There's a repo called SD scripts on GitHub I'll link to in a minute, another one called Cog SDXL. They both have pros and cons, but you can basically send these images into these things and they will spit out a lora, because I helpfully explained Lora earlier, it's basically a diff between the weights in the model and the updated weights in the model. So the nice thing about a lower file is like the actual weights of SDXL might be 10gb or something, the lower can be five megs, and so you almost fit it on a floppy disk. But these things like you can move them around easily and you can make lots of them, and you can easily swap them in and out of a running model. Fine tuning text is a bit more complicated. So if you've got like a bunch of documents, web pages, pdfs, papers, whatever, you want, to start by converting them into plain text or markdown. All of these models are actually markdown machines. Like, it seems like everyone turned their data into markdown because it's a nice convenient format.
Luke Marsden [00:29:28]: You might see chat, GPT, spit out some half formed markdown sometimes, and then it will quickly get turned into bold text or something. Plain text is fine too, because plain text is markdown. Then we put together this pipeline for doing fine tuning on text, where you then as part of your fine tuning pipeline, you use a large language model to convert the text into question answer pairs. And why would you do that? It's because of this thing I was talking about earlier, this instruction tuning idea, the plain text that you've got from Wikipedia or internal business document docx or whatever, is suitable for a completion style model. But you presumably most people don't want completion style models. They want instruction style models where you can ask the model questions about the document. And if you want to fine tune a model to give you answers to questions, then you should give it question answer pairs as the training set. And so it would be very tedious to manually turn your word documents into question answer pairs, right? But what you can do is you can use a large language model to do that job for you.
Luke Marsden [00:30:42]: So you can give it a prompt. And the prompt is quite funny because it's like imagine you're a professor setting a test. You're given this content. I want you to ask questions about the give me questions and answers. So we then can take those qa pairs that are in a JSon L file. JSon L file is just line delimited JSon. It says question colon, answer colon in the file. And then you can feed that into Axolotl, which understands this as a training data input type.
Luke Marsden [00:31:21]: And that will spit out a lora. The LLM fine tuned lowers are a bit bigger, but they're still, I think a few hundred megs, but they're still quite easy to move around. And then you can do inference with Lora on both text or image models. So this is like given a model and a prompt, you can additionally plug in this Lora file. And you know the bit in the matrix where they plug something into his brain and now he knows kung fu. I feel like it's a bit like that for the language model. It's like you plug in the lower file and now it knows about, sorry, diggers. Diggers, exactly, diggers.
Luke Marsden [00:31:58]: Or like big co's HR policies or whatever. So that's kind of interesting. And then just coming back to this, then there's kind of this bin packing problem of like, okay, I want to run some fine tuning workloads on my gpus. I want to run some inference with on a base model, and I also want to run some inference with the loweras, how do I pack those in? So that's something that we've been working on as well. Then there's also this topic of retrieval, augmented generation, which I won't go into in lots of detail, but llama index is a good place to look at information on this. And it's basically this idea that you take data that you've got, you split it up into little chunks, and you index the chunks in a way that turns them into vectors. So it's called an embedding model, and it basically maps the sentence, or whatever, let's call it a sentence, it maps the sentence into a high dimensional space that kind of says, oh, it's going to be close to other sentences that are about similar things. And you can do fun algebra in the embedding space, which is like, I can never get this example right, but it's something like king plus woman equals queen, because you can kind of move around the embedding space, but it's like actually 1024 dimensions, so you have to squash it down if you want to visualize it.
Luke Marsden [00:33:32]: So anyway, you can put all your sentences into this index, and then the user will send a query to the system. The query will look up similar documents in the index, and then it will just include those documents in the prompt before sending them to the language model. And this is quite powerful because you don't need to do the slower fine tuning step, and you can give the user, well, the language model is more accurate when you do rag in this way, and it also makes it easy to give the language model access to live information. So there's that. So I'm going to show you a few quick demos and then I'll wrap up, because I'm over time, everything I'm going to show you and these demos are just on slides because I was lazy. But everything I'm going to show you can be done with Axolotl, CoG, SDXL and SD scripts, or you can go to our GitHub and we put a UI around it, basically including the ability to do the bin packing onto gpus that I talked about. So the first use case is brainstorming business ideas. So this is just a use case or a demo of text inference.
Luke Marsden [00:34:49]: So I asked Mistral the other day, the US Patriot act and EU telco regulation is one reason why EU telcos might want to run llms on Prem. What are some other compelling business reasons that high value customers might, blah blah blah, want to run LLMs on Prem? And it gave them to me. And I was like, okay, well, that's useful. Now give me the top ten industry location pairs for this strategy, along with reasons why they might buy something. And it does it. It gives me ten locations, it says, but it works basically. It gives you accurate, good ideas about things that you might want to try. So text inference, the most simple version of this thing is useful on its own.
Luke Marsden [00:35:29]: And you can see this from the fact that Chat GPT has been massively successful and people use it all the time. Image inference is just like simple. Come up with a logo for my VR headset company. It's like, okay, well, they're fine, they're not great, but they're good enough. Good enough to start something, probably, or good enough to come up with some ideas if you're a designer. So, yeah, you can just type this in and type this in here. And, yeah, here it is. The next example is text fine tuning.
Luke Marsden [00:36:01]: So an example I really like, which actually shows off text fine tuning quite nicely, is I want to understand and interrogate brand new scientific papers that have just come out today. Chat GPT can't know about them yet because it was trained like this. Knowledge cut off was April 23 or something. Or even if it was like last month, it's still not going to know about the latest scientific knowledge in the world. So let's go to archive.org. Someone shout out one of these areas that they want to learn something about. Quantitative finance. Quantitative finance.
Luke Marsden [00:36:38]: Okay, let's look at new quantitative finance papers. Which one would you like? I'm just going to give you this one. Yeah. Economic forces in stock returns. Okay, so let's go here. Oh, yeah, there we go. A green one. So we can fine tune some text.
Luke Marsden [00:36:59]: Just paste in the paper. It's going to attach it here. It's now going to download and extract the text from that paper. I hope that worked. Yes, it did. A few. Okay. And then it's going to chunk up that text into question answer pairs.
Luke Marsden [00:37:18]: Well, it's going to chunk up the text and then pass it into the large language model to generate the question answer pairs. And then it's going to fine tune the language model. So here's one I made earlier. This is about galaxy clusters. I don't know anything about this topic, but it's about radio galaxies. So I gave it this paper and then I said, just what frequencies were used in the study? It knows. And then I was like, what the hell are redtail radio galaxies? And then it probably leans on its other internal knowledge about what those are. So I feel like that's actually quite useful because now I can learn about things in papers that I wouldn't be able to understand otherwise, because I can ask the model to explain the basics to me as well.
Luke Marsden [00:38:02]: And then there's an image fine tuning model, which is like. Or image fine tuning demo, which is like, okay, I'm a photographer and I want to make variations of my own photographs. So suppose hypothetically, that I took these photographs. I actually did take the first two. And then I want more boats in the San Francisco like city skyline. And you see how you can probably just about see it. The image looks almost exactly the same. It's quite nice.
Luke Marsden [00:38:28]: So that's like the original image. And then this is basically kind of a prettier version of the same image with more boats, because I asked for boats. And then there's another one which is like, oh, now I want to see the San Francisco bridge in the dark. It does a reasonably good job of it. So as a photographer, I might want to be able to do this sort of creative imagining on my own images that I can upload anyway. So this is nearly the end of my talk, because, I'm sorry, I'm over time. There's lots of interesting things that are coming down the pipe in terms of multimodal models. So GPT four vision already exists.
Luke Marsden [00:39:06]: You can give it an image and it will be able to read the image. Like the hacking example I gave earlier. There's open source versions of those, I think there's one called backlava, and then there's all the fun stuff you can do with things like control nets in SDXL to make funny looking images, but I think those are going to be interesting to expose as well. And then there's the whisper models to do like speech, text and so on. And I already talked about this a bit, but I think code is a really interesting use case, especially with local fine tuned llms on a private data set. Voice obviously is huge. And yeah, this is just 10 seconds on Helix. So we started this new project in December.
Luke Marsden [00:39:52]: I've done a couple of other startups before. We're aiming to bootstrap this one, so we need some customers. Google's we have no moat paper was really interesting because it basically says we don't have an advantage against the open source stuff. And then you see the open source stuff getting faster, better, faster and faster. So it's like kind of interesting timing to have another go. There's this rapid acceleration of new open source models can fine tune them on your own hardware, so. Yeah, that's kind of why we're doing it. Thanks very much.