-
Notifications
You must be signed in to change notification settings - Fork 231
Process unstructured data sources #251
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
A reference: https://hal.science/hal-03430826/document |
This comment was marked as outdated.
This comment was marked as outdated.
Interested in the Project Idea... |
Please also check: https://github.com/cve-search/git-vuln-finder |
I guess the process of change logs of Apache mailing list can be automated using OpenAI' API or other open source LLMs, where we scrape the data using Selenium, feed into LLM, get the output as json format and then update the database accordingly. What is your view on that @pombredanne . #218 Can also be implemented. |
Automating the extraction of valuable information from Apache mailing list changelogs using OpenAI’s API and other tools is a great initiative and I think for the unstructured data we can focus primarily onto the Dataset for feature Engineering and classified into diverse group Model Training: Fine-tune the selected model on a prepared dataset of CVEs in code. This will help the model learn to identify vulnerabilities in the unstructured data..... As well as we can use LoRA for the model to train Vulnerability Detection: Use the trained model to parse through the unstructured data and identify potential vulnerabilities. This could involve using NLP techniques to understand the vulnerability descriptions and infer the vulnerable package name and versions. Most Important Parameter to be checked is this Information Extraction: This is the process of automatically extracting structured information from unstructured text data. |
These contain valuable data nuggets among an ocean of junk and we need to be able to find the good things there.
Some sources are:
We can either automate it all, but that's going to be super difficult, or rather start to craft a curation queue and parse as much as we can to make it easy to curate by humans
... and progressively also improve some mini AI and classification to help further automate the work.
The text was updated successfully, but these errors were encountered: