Building Arivu: CLI/library that normalizes fetch/search across a bunch of sources (arXiv, PubMed, HN, GitHub, Reddit, YouTube transcripts, RSS, web pages…).
I use it as a context fetcher i.e grab an abstract/transcript/thread as clean text/JSON, pipe it into summaries or scripts.
Also runs as an MCP server (experimental), so tools like Claude Desktop or CLI assistants can call the connectors directly.
If we are already comfortable with our enterprise chatgpt subscription, how might this be of value. Given that it does RAG, tool calling, has all the SSO stuff/collab? Or are we not the target customer.
Just curious. Especially with both OpenAI and Anthropic really also outpacing startups in release cadence unlike previous cycles.
Guessing your selling point is any model no locking (Assuming we are happy with the privacy SOC 2 etc guarantees on enterprise contracts here)
I built a prototype using native messaging (the same way apps password managers interact with browsers and drive actions with pure js).
I have a lot of actions done but not full there yet. Essentially the goal is to use a cli or an external desktop app to drive your already logged‑in Chrome profile without navigator.webdriver, or enabling --remote‑debugging‑port. In all my testing never got flagged with captcha/bot protect. The cli can interact with LLMs, local file system(despite opfs this is easier).
Extenion just executes pure js for extraction, navigation. Ive gotten basic google searching, image downloading working. Looking at complex form interactions.
native messaging is a huge headache to set up reliably across all the OSs and (often headless) chrome setups in our experience, that's why we've avoided it.
Just using some remote endpoint message bus service is an easier solution, or something like ElectricSQL/RxDB/Replicache/etc.
We also can't really use in-page JS for much because it's easily detected by bot-blockers, though isolated worlds help in CDP.
Managed to make long dictations even >10mins appear in < 2seconds by pushing what is possible with current STT models.
All processing done locally with 0 network calls.
reply