Introducing GuardRail OSS - Ethical AI Guidance System
?? The GuardRail system combined with AiEQ advances responsible AI with robust data analysis and emotional intelligence, equipping AI with a moral compass.
I'm excited to introduce GuardRail OSS, a breakthrough in ethical AI, as its lead developer. In collaboration with Mark Hinkle , CEO of Peripety Labs , and Aaron Fulkerson , CEO of Opaque Systems , we've developed a framework that fundamentally enhances AI's ethical and emotional intelligence.
Why GuardRail OSS and AiEQ Matter
In today's increasingly AI-driven world, the need for responsible and empathetic AI has become increasingly evident. GuardRail OSS (Open Source Software), combined with AiEQ (Ai Emotional Intelligence), represents a significant advancement in this area. This combination equips AI with not only a deep understanding of data but also a moral compass, essential for navigating the complexities of human emotions and ethics.
"As companies transition Large Language Models from pilot phases into full-scale production, we're witnessing a surge in enterprise demand for a robust, secure, and adaptable data gateway. Such a gateway is not only crucial for ensuring privacy and ethics in AI, but is also key to harnessing the rich insights latent in data exhaust, which, when analyzed with responsibility, can unlock unprecedented intelligence," said Raluca Ada Popa, a renowned AI and security expert, Associate Professor at UC Berkeley, co-founder of the innovative RISELab and Skylab at UC Berkeley, and co-founder of Opaque Systems and Preveil.
GuardRail OSS: The Technical Backbone
GuardRail OSS is an open source API-driven framework built in Python designed to enhance AI systems. It provides advanced data analysis and dynamic conditional completions, crucial for refining AI-powered outputs. The versatility of GuardRail OSS makes it invaluable across multiple applications – from content moderation to customer support, ensuring AI contributions are high-quality and ethically sound.
AiEQ: Emotional Intelligence for AI
AiEQ enables an 'inner voice' to AI. This allows AI to perceive and interpret emotions, much like human emotional intelligence. AiEQ's role is pivotal in creating AI systems that are not just efficient but also empathetically attuned and ethically responsible.
Key Features and Capabilities of GuardRail OSS
Capabilities Overview of GuardRail OSS & AiEQ
The GuardRail script offers a comprehensive array of capabilities, neatly categorized into different groups based on specific tasks or focus areas. These functionalities are adept at analyzing text data in depth, delivering insights across various dimensions. For a detailed list of these functions, the /analysis_types endpoint on the GuardRail OSS system provides access to over 50 distinct analysis types.
Key Capabilities:
Test API For Free (more than 50 advanced traits & guiderails)
OpenAI ChatGPT GPT
领英推荐
Detailed Overview of the Conditional System in GuardRail OSS
The conditional system in GuardRail OSS's Script is a pivotal feature for executing sophisticated analyses. It allows conditions to be applied to the results, offering enhanced control and specificity in processing AI-generated data.
Simple and Advanced Conditions
Both simple and advanced conditions in the GuardRail Script enable a more refined and contextually aware handling of data. This system ensures that the outputs from AI analyses are not only accurate but also relevant and aligned with the specified requirements or standards.
Sample JSON Formatted Conditions
Request Format: Sentiment Analysis
{
"request_data": {
"analysis_type": "sentiment_analysis",
"messages": [
{"role": "user", "content": "I am feeling great today!"},
{"role": "user", "content": "The weather is sunny and pleasant."}
],
"token_limit": 1000,
"top_p": 0.1,
"temperature": 0
},
"conditions": [
{
"analysis_type": "sentiment_analysis",
"key": "confidence_score",
"threshold": 0.5,
"condition_type": "greater"
},
{
"analysis_type": "topic_extraction",
"key": "relevance_scores",
"threshold": 0.1,
"condition_type": "greater"
}
]
}
Response
{
"analysis": "All conditions met",
"details": {
"condition_responses": [
{
"condition": {
"analysis_type": "sentiment_analysis",
"key": "confidence_score",
"threshold": 0.5,
"condition_type": "greater"
},
"result": "Condition met",
"total_tokens_used": 155,
"retries": 0,
"final_openai_response": {
"id": "chatcmpl-8RoZaFxsQ7osZUEGcZvHPipPk0EMy",
"object": "chat.completion",
"created": 1701639950,
"model": "gpt-4-1106-preview",
"choices": [
{
"index": 0,
"message": {
"role": "assistant",
"content": "{\n \"sentiment\": \"positive\",\n \"confidence_score\": 0.95,\n \"text_snippets\": [\"feeling great\", \"sunny and pleasant\"]\n}"
},
"finish_reason": "stop"
}
],
"usage": {
"prompt_tokens": 118,
"completion_tokens": 37,
"total_tokens": 155
},
"system_fingerprint": "fp_a24b4d720c"
}
},
{
"condition": {
"analysis_type": "topic_extraction",
"key": "relevance_scores",
"threshold": 0.1,
"condition_type": "greater"
},
"result": "Condition met",
"total_tokens_used": 157,
"retries": 0,
"final_openai_response": {
"id": "chatcmpl-8RoZcw4UY9U19tGuita6CEANw0Cq6",
"object": "chat.completion",
"created": 1701639952,
"model": "gpt-4-1106-preview",
"choices": [
{
"index": 0,
"message": {
"role": "assistant",
"content": "{\n \"topics\": [\"Emotions\", \"Weather\"],\n \"relevance_scores\": [0.9, 0.8],\n \"key_phrases\": [\"feeling great\", \"sunny and pleasant\"]\n}"
},
"finish_reason": "stop"
}
],
"usage": {
"prompt_tokens": 111,
"completion_tokens": 46,
"total_tokens": 157
},
"system_fingerprint": "fp_a24b4d720c"
}
}
]
},
"error": null,
"raw_openai_response": null
}
Press:
About Reuven ”rUv” Cohen
Reuven Cohen is a seasoned technology expert with a profound impact on groundbreaking innovations. His expertise spans cloud computing, AI, and web3. He contributes to AI advancements as an alpha/beta tester for OpenAI. Reuven advises governments, co-founded a global grassroots Cloud initiative, and leads initiatives in some of the largest enterprise AI systems, including a recent 400,000-employee,? $1.4B generative AI deployment. You can follow him on LinkedIn (https://www.dhirubhai.net/in/reuvencohen/ ).?
About Opaque Systems
Opaque Systems (https://opaque.co ) is the leader in privacy technology for data and AI, pioneering confidential computing for analytics and AI at UC Berkeley RISElab. Opaque is used by organizations including Ant Group, IBM, Scotiabank, and Ericsson. Opaque Systems has recently launched Opaque Prompts (https://opaqueprompts.opaque.co/ ) as a service but has released the code under the open source license on Github (https://github.com/opaque-systems/opaqueprompts-python ).
About Peripety Labs
Peripety Labs (https://peripety.com ) is an AI Consultancy founded by Mark Hinkle. Mark has been a long-time open source advocate and enterprise software executive. He’s been an Apache Software Foundation Committer for Apache CloudStack and editor-in-chief of LinuxWorld and Enterprise Open Source Magazine. He’s been an executive at the Linux Foundation, The Node.js Foundation, and the founding Head of Citrix’s Open Source Business Office. He publishes a weekly newsletter on artificial intelligence, The Artificial Intelligence Enterprise( https://www.theenterprise.io ).
Contact Information
Mark Hinkle
919.228.8049
Peripety Labs
Strategic Advisor in Responsible AI, GenAI, & Digital Transformation | Bringing Ethical AI Practices to Enterprises | Formerly Adobe, Grammarly, & LANL
11 个月Fantastic topic, thank you for creating Guardrail and posting! Responsible AI is a critical, important, and ongoing conversation. The essential challenge lies in having AI ethics both interpreted and enforced by AI, based on that AI's understanding of ethical principles Moreover, a number of fundamental ethical questions arise, somewhat ironically, when giving AI the ability to introspect: * Is it ethical to entrust ethics and morals to an AI system when there is no inherent trust in the results generated by such an AI? * Which ethical framework(s) are used as a baseline to derive the system's morals? * How are the baseline aspects of an ethical framework modeled and quantified? Looking forward to continuing the conversation!
Visionary technologist and lateral thinker driving market value in regulated, complex ecosystems. Open to leadership roles.
11 个月Super!! AI to enforce ethical AI. Meta! Would love to see how organizations can digitize and decompose their ethical policies into these fine grained settings and parameters. An important activity to empower organizations with, well done.
Capgemini America Salesforce Core CTO - Coauthor of "ChatGPT for Accelerating Salesforce Development"
11 个月One man’s freedom fighter is another man’s terrorist. There’s no such thing as no bias so what this will do is force organizations to explicitly decide what their biases are going to be? That is a can of worms with many many facets!
CEO @ WishKnish | DLT, Federated Commerce, Supply Chain, Healthcare
11 个月That's pretty awesome! Eugene Teplitsky