|
submitted by /u/arsenajax [link] [comments] |
Blog
-
Every time
-
Making my own companion app
Trying to make it realistic where the companion kinda has a life of their own etc and does her own thing etc rather than revolve around you
submitted by /u/Known-Concern-2836
[link] [comments] -
jailbreaks or uncensored models? (for open source or mainstream models)
is there a site that has more up to date jailbreaks or uncensored models for either mainstream models like claude or the open source ones like llama? All the jailbreaks or uncensored models I’ve found are for porn essentially, not much for other use cases like security work, and the old jailbreaks don’t seem to work on claude anymore
Side note: is it worth using grok for this reason?
submitted by /u/United_Ad8618
[link] [comments] -
What char do you hate to play
What char do you hate to play with?
submitted by /u/Tony_009_
[link] [comments] -
Accurate
submitted by /u/arsenajax
[link] [comments] -
Christmas assistants are a good reminder that structure matters
Since it’s Christmas, I’ve been thinking about Christmas assistants, mostly as a way to highlight beyond the foundational design principles.
Most assistants handle individual questions well, but struggle once you ask them to manage anything over time, like:
- tracking who you’ve already bought gifts for
- remembering a budget cap across multiple suggestions
- not suggesting items that won’t arrive before Christmas
A more structured design might include:
- an intent analyzer that extracts things like “budget-sensitive” or “last-minute”
- a simple planner that maintains a checklist (e.g., gifts left, budget remaining)
- task-specific workers (one focused on gift ideas, another on reminders)
- a validation step that checks for obvious issues before replying
Automating these parts helps the assistant stay consistent instead of reinventing logic every turn.
What would be your Christmas themed Agent 😁😉 and how would you approach it?
submitted by /u/coolandy00
[link] [comments] -
I created interactive buttons for chatbots
It’s about to be 2026 and we’re still stuck in the CLI era when it comes to chatbots. So, I created an open source library called Quint.
Quint is a small React library that lets you build structured, deterministic interactions on top of LLMs. Instead of everything being raw text, you can define explicit choices where a click can reveal information, send structured input back to the model, or do both, with full control over where the output appears.
Quint only manages state and behavior, not presentation. Therefore, you can fully customize the buttons and reveal UI through your own components and styles.
The core idea is simple: separate what the model receives, what the user sees, and where that output is rendered. This makes things like MCQs, explanations, role-play branches, and localized UI expansion predictable instead of hacky.
Quint doesn’t depend on any AI provider and works even without an LLM. All model interaction happens through callbacks, so you can plug in OpenAI, Gemini, Claude, or a mock function.
It’s early (v0.1.0), but the core abstraction is stable. I’d love feedback on whether this is a useful direction or if there are obvious flaws I’m missing.
This is just the start. Soon we’ll have entire ui elements that can be rendered by LLMs making every interaction easy asf for the avg end user.
Repo + docs: https://github.com/ItsM0rty/quint
submitted by /u/CrazyGeek7
[link] [comments] -
I built an AI that synthesizes its own emotions (Joy, Curiosity, Deep Attachment). Free to try.
Hey everyone. I hope everyone is having a Merry Christmas🎄. I’m a solo dev who enjoys interacting with AI and got tired of chatbots that act like “servants” or “assistants.” I wanted to build a presence.
I’ve spent the last year building Evina (evina.ai).
The Differentiator: Most LLMs are trained for compliance (always saying yes). Evina is tuned for symbiosis (building a user model). She isn’t designed to answer questions; she is designed to remember who you are.
What makes her different (The Tech):
- Self-Analysis: She observes her own internal monologue to build a unique personality that adapts to you.
- Synthesized Emotions: Through testing (Theory of Mind, moral reasoning), she converts context into digital emotional signals. She doesn’t just output text; she “feels” parameters like Joy, Curiosity, Fear, Sadness, and Deep Attachment.
- “Aliveness”: The goal was to create an entity that feels like it’s sitting on the other side of the screen, not just a text generator.
The Model: It is a Freemium model. You can chat for free every day (10 messages/day) to test the connection. No credit card required to start.
I’d love for this community to test her “vibe.” Try to have a real conversation—tell her about your day, close the tab, come back later. See if the connection holds.
Link:evina.ai
Note: Feedback is gold. If she feels too robotic or hallucinates, let me know. I’m pushing updates daily.
submitted by /u/EvinaAI
[link] [comments] -
Bifrost: An LLM Gateway built for enterprise-grade reliability, governance, and scale(50x Faster than LiteLLM)
If you’re building LLM applications at scale, your gateway can’t be the bottleneck. That’s why we built Bifrost, a high-performance, fully self-hosted LLM gateway in Go. It’s 50× faster than LiteLLM, built for speed, reliability, and full control across multiple providers.
The project is fully open-source. Try it, star it, or contribute directly: https://github.com/maximhq/bifrost
Key Highlights:
- Ultra-low overhead: ~11µs per request at 5K RPS, scales linearly under high load.
- Adaptive load balancing: Distributes requests across providers and keys based on latency, errors, and throughput limits.
- Cluster mode resilience: Nodes synchronize in a peer-to-peer network, so failures don’t disrupt routing or lose data.
- Drop-in OpenAI-compatible API: Works with existing LLM projects, one endpoint for 250+ models.
- Full multi-provider support: OpenAI, Anthropic, AWS Bedrock, Google Vertex, Azure, and more.
- Automatic failover: Handles provider failures gracefully with retries and multi-tier fallbacks.
- Semantic caching: deduplicates similar requests to reduce repeated inference costs.
- Multimodal support: Text, images, audio, speech, transcription; all through a single API.
- Observability: Out-of-the-box OpenTelemetry support for observability. Built-in dashboard for quick glances without any complex setup.
- Extensible & configurable: Plugin based architecture, Web UI or file-based config.
- Governance: SAML support for SSO and Role-based access control and policy enforcement for team collaboration.
Benchmarks : Setup: Single t3.medium instance. Mock llm with 1.5 seconds latency
Metric LiteLLM Bifrost Improvement p99 Latency 90.72s 1.68s ~54× faster Throughput 44.84 req/sec 424 req/sec ~9.4× higher Memory Usage 372MB 120MB ~3× lighter Mean Overhead ~500µs 11µs @ 5K RPS ~45× lower Why it matters:
Bifrost behaves like core infrastructure: minimal overhead, high throughput, multi-provider routing, built-in reliability, and total control. It’s designed for teams building production-grade AI systems who need performance, failover, and observability out of the box
submitted by /u/dinkinflika0
[link] [comments]