A chatbot is a software application designed to simulate human conversation. It interacts with users through messaging platforms, websites, or mobile apps, using pre-set scripts or artificial intelligence technologies to understand queries and provide responses. Chatbots are often used for customer service, information retrieval, or as virtual assistants.
Open-weight large language models (LLMs) are as effective as closed-weight systems for structuring radiology reports while ensuring patient data privacy through local implementation.
Researchers from the University of Basel reveal that trust in AI chatbots is shaped by integrity and competence, with personalization enhancing user engagement but not trust.
Amazon Nova revolutionizes foundation models with state-of-the-art generative AI capabilities, delivering unparalleled performance for multimodal tasks, document analysis, and creative content generation while prioritizing customization and ethical AI use.
Researchers from Cambridge developed a method to detect "bullshit" in ChatGPT’s outputs, revealing striking parallels between AI-generated text and uninformative human discourse like political language and "bullshit jobs."
Researchers introduce AMD OLMo, an open-source language model with 1 billion parameters, trained using 1.3 trillion tokens on AMD GPUs to push the boundaries of AI, enabling improved reasoning, instruction-following, and ethical alignment in AI systems.
A new watermarking method, SynthID-Text, enables clear identification of AI-generated text while preserving quality and efficiency.
The G7 Toolkit for Artificial Intelligence in the Public Sector outlines strategies for ethical, secure, and effective AI deployment in governments, emphasizing human rights and transparency. It includes case studies and best practices to guide responsible AI adoption globally.
Researchers created 1.5 million AI-generated material narratives to address biases in materials science, enabling a more balanced exploration of solid-state materials for new applications.
Researchers introduced LOLA, a massively multilingual LLM utilizing a sparse Mixture-of-Experts architecture, outperforming larger models on multilingual tasks with efficiency and scalability.
Researchers introduce a new method to efficiently differentiate large language models (LLMs) in a black-box setting using fewer than 20 benign binary questions, improving accuracy and transparency in AI audits.
Researchers propose revisions to trust models, highlighting the complexities introduced by generative AI chatbots and the critical role of developers and training data.
Generative chatbots significantly increase the formation and persistence of false memories during simulated crime witness interviews, raising ethical concerns about their use in sensitive contexts.
Aleph Alpha has introduced the Pharia-1-LLM-7B models, optimized for concise, multilingual responses with domain-specific applications in automotive and engineering. The models include safety features and are available for non-commercial research.
Researchers explored using transfer learning to improve chatbot models for customer service across various industries, showing significant performance boosts, particularly in data-scarce areas. The study demonstrated successful deployment on physical robots like Softbank's Pepper and Temi.
A study published in Future Internet explored the use of multimodal large language models (MLLMs) for emotion recognition from videos. The researchers combined visual and acoustic data to test MLLMs in a zero-shot learning setting, finding that MLLMs excelled in recognizing emotions with intensity deviations, though they did not outperform state-of-the-art models on the Hume-Reaction benchmark.
Researchers recently introduced the CHEW dataset to evaluate large language models' (LLMs) ability to understand and generate timelines of entities and events based on Wikipedia revisions. By testing models like Llama and Mistral, the study demonstrated improvements in tracking information changes over time, thereby addressing the common issue of temporal misalignment in LLMs.
Researchers explored the potential of large language models (LLMs) like GPT-4 and Claude 2 for automated essay scoring (AES), showing that these AI systems offer reliable and valid scoring comparable to human raters. The study underscores the promise of LLMs in educational technology, while highlighting the need for further refinement and ethical considerations.
Researchers introduced "Chameleon," a mixed-modal foundation model designed to seamlessly integrate text and images using an early-fusion token-based method. The model demonstrated superior performance in tasks such as visual question answering and image captioning, setting new standards for multimodal AI and offering broad applications in content creation, interactive systems, and data analysis.
A recent Meta Research article explored semantic drift in large language models (LLMs), revealing that initial accuracy in text generation declines over time. Researchers introduced the "semantic drift score" to measure this effect and tested strategies like early stopping and resampling to maintain factual accuracy, showing significant improvements in the reliability of AI-generated content.
Researchers explored whether ChatGPT-4's personality traits can be assessed and influenced by user interactions, aiming to enhance human-computer interaction. Using Big Five and MBTI frameworks, they demonstrated that ChatGPT-4 exhibits measurable personality traits, which can be shifted through targeted prompting, showing potential for personalized AI applications.
Terms
While we only use edited and approved content for Azthena
answers, it may on occasions provide incorrect responses.
Please confirm any data provided with the related suppliers or
authors. We do not provide medical advice, if you search for
medical information you must always consult a medical
professional before acting on any information provided.
Your questions, but not your email details will be shared with
OpenAI and retained for 30 days in accordance with their
privacy principles.
Please do not ask questions that use sensitive or confidential
information.
Read the full Terms & Conditions.