Usage

Below, a simple example of how to run SciWatch for monitoring LLM, document augmentation, and Active Learning papers, and blogs.

Setup senders

See Senders.

Create a configuration file

Lets suppose we named the file scrapping_config.toml.

title = "LLM & AL Watch" # Will be used as email title

end_date = "now" # will search content up to now (exec. time)
time_delta = "02:00:00:00" # will look for content up to two days ago

recipients = ["aghiles.ahmed.azzoug@gmail.com"]

# define your queries
[[query]]
title = "LLM" # LLM query
raw_content = """intitle:(GPT* OR LLM* OR prompt* OR "Large language models"~2) AND incontent:(survey OR review OR evaluation* OR benchmark* OR optimization*)"""

[[query]]
title = "AL" # Active Learning on VRD (or benchmarks/surveys)
raw_content = """intitle:("active learning") AND incontent:(VRD OR documents OR survey* OR benchmark*)"""

# define your sources
[[source]]
type = "arxiv" # check for Computer Science papers on Arxiv
use_abstract_as_content = true
search_topic = "cs"
max_documents = 200

[[source]]
type = "openai_blog" # check for latest blogs on OpenAI blog (mainly for GPT updates)
max_documents = 20

For more information about query syntax see Query syntax.

Run the watcher

from sci_watch.sci_watcher import SciWatcher

watcher = SciWatcher.from_toml("scrapping_config.toml")

watcher.exec() # if some relevant content is retrieved, recipients will receive an Email

You might get an email like this:

Email sample

Tip

You can run your code through Crontab and receive daily (or weekly) updates!