Replies: 6 comments 5 replies
-
|
Hi! Thanks for kicking off this discussion. As a user, I can say unequivocally that flexibility in how the AI features are implemented is most important. For someone with unlimited RAM and CPU resources, using llama.cpp might be perfect, but I know plenty of situations where you’d want the option to use cloud-based LLMs or LLMs deployed on a local network. A private mode can (and perhaps should) be the default, but all the other options also deserve to exist. |
Beta Was this translation helpful? Give feedback.
-
|
I’m thinking model risk assessment could be integrated. It seems, on a personal level, that many (not necessarily most and probably not all) current deployments lack what should be appropriate safety / security / risk averse controls to safeguard end-users. How does or will this infrastructure be implemented with risk in mind? I could possibly ask this and other questions associated with that subject. |
Beta Was this translation helpful? Give feedback.
-
The model space is constantly evolving. Best would be to define a set of test cases and evaluate them regularly. Best: Make test cases such that they are fed with random data each time so no one can train their model to one specific test set. Then let users evaluate their models voluntarily and send in their evaluations with a click of a button. Take care to get the model version and setup. This way you will quickly build up a database. Publish it on github. If it gets featured as standard evaluation this will get more exposure to the ODE.
Better hardware, caching of previous results if they are correct. In general: Prefer small and specialised models.
What are you going to achieve with this question?
Hallucinations in LLM are not a bug, they are a feature. Do not use LLM to analyse data directly. Data extraction is possible but needs to be cross checked. The usual way to go is to use the LLM to understand what the user wants and to propose solutions. The actual data wrangling needs to be done by deterministic algorithms that can be checked for correctness. -> You may want to offer fine tuned small local LLM for that.
I guess next to quality of response mostly response time and costs.
First get a good overview on the model performance (See question 1). Then this question maybe has a solution. If you need professional help with any of these topics for implementation, I know a decent company. |
Beta Was this translation helpful? Give feedback.
-
|
Some thoughts about these questions
Yes, if LLMs prompting is behind the scenes. Data prompting require skills much above non-technical people. Non-technical people should be actively assisted by application. There are typical operations that could/should/would be implemented to help them, but they shoud use it with predefined prompts that return structured outputs. Modern LLMs help a lot to analyze columns names, auto-generate documentations on data, detect semantic types and e.t.c
Yes, but it's very expensive to implement it in local only mode. It may require iterative agentic approach. There are several startups that implement such assistance but almost all of them use cloud-based or other network based LLMs.
Yes, but LLMs are secondary to rule based validation. Tools like Great expecations or Soda data quality are examples of the data validation frameworks without AI. LLM could generate these rules to speed-up rules definitions.
It could be done even without LLMs by providing proper guide. LLMs could be used to validate names or predefined rules, but it's not nessesary. Anyway it's not about prompting by user, it's about predefined prompts and buttons like "Suggest column names using LLM" or similar.
Sure, but these best practices should be defined first
It's a next step after detection of errors aand inconsistency using less resource consuming methods.
User AI and data and metadata education is not the same as a tool that helps to impove metadata and data quality. |
Beta Was this translation helpful? Give feedback.
-
|
Which models are more suitable for the tasks we have in mind? - The ones which are purposefully trained for that. |
Beta Was this translation helpful? Give feedback.
-
|
3 months later and things move quickly. Did you have a look at the IBM Granit models lately? They feel like GPT3 but run on a few GB locally. I don't need a model that has tons of general knowledge. I need 1 model that understands text, 1 model that can code python and 1 that analyses images. Each one just needs a few GB and could run locally. The first one can call the others and hand over the problem. Also, I recently discovered https://github.com/sst/opencode and I am impressed. Yes, it is in the terminal and the selling point is code. But that is far from all. You can add models that are not coding-specific and then have them munch through your text data with ease. You can use coding models (grok code 1 fast that comes with it at the moment for free is pretty decent, but you can add a different if you like) to let them write python scripts that analyse your data. Just follow test driven development working with it and it is such a powerful tool. You can define assistents ("agents") with markdown that will take care of the work. And everything is in the directory you work from, in markdown text. ODE could build on that. Add it as backend. Allow it to call ODE as tool and vice versa. |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Local LLMs have been already integrated into
mainbranch and the beta feature will be released soon. Since the technical integration has been completed, it is time to start thinking about performance, quality, use cases, etc.Context
The current setup is using llama.cpp server so we have the flexibility to use any model we want as well as customize it.
The current landscape of models and use cases is huge so there are a couple of questions that requires research or consulting from someone expert in the field. The goal of the integration is to explore LLM capabilities when working with data and metadata or in other works How useful are LLMs whe working with data?.
Use cases/Hypothesis
Some questions or unknowns:
Beta Was this translation helpful? Give feedback.
All reactions