SenseMirage "is a creative platform that includes SenseMirage's self-developed AIGC large model and convenient LoRA training capabilities, as well as providing third-party community open-source models to accelerate inference. It provides creators with more convenient and comprehensive content production and creation tools. SenseMirage not only integrates SenseMirage's self-developed text and image generation model with over 1 billion parameters, but also supports inference acceleration optimization for non self-developed models. Combined with self-developed models and training capabilities, SenseMirage eliminates the need for localized deployment processes and efficiently generates more diverse content based on self trained LoRA models. Gewu "is a 3D content generation platform developed by SenseTime based on neural radiation field technology (NeRF). It has the ability to finely generate 3D objects and can restore complex geometric structures, textures, materials, gloss, and other details of objects, achieving real-time high fidelity rendering. Mingmou "is a data annotation platform based on Shangtang's self-developed large model, which has more than 10 universal large models and industry-specific large models built-in. It supports intelligent annotation for 2D classification, detection, and 3D detection of various scenarios such as intelligent driving, intelligent transportation, and smart cities; Compared with traditional manual annotation and small model annotation modes, it has the core advantages of good annotation effect, high efficiency, and low cost. Through standard APIs, the underlying scheduling system of SenseTime provides massive computing power to support large-scale annotation needs. 'Consultation' is an artificial intelligence language model developed by SenseTime Technology based on natural language processing technology, which has extraordinary language understanding and generation capabilities, integrating technology and humanities. As an efficient chat assistant, it can quickly solve complex problems, provide customized suggestions, and assist in creating top-notch text, with the characteristic of continuous learning and evolution. Qiongyu "is a 3D content generation platform developed by SenseTime based on neural radiation field technology (NeRF), which has the ability to generate large-scale spatial reconstruction at the city level. The generated 3D content can be re edited and re created on the platform, meeting the needs of industries such as film and television creation, architectural design, product marketing, and digital twin management and operation through the production of massive high-precision digital assets. SenseTime "Shadow Shadow" is an application platform launched by SenseTime Technology, with digital character generation technology as its core, based on SenseTime's various AI generation capabilities, including text generation, speech generation, action generation, image generation, NeRF, etc. 'Shadow Shadow' can help users break free from the constraints of professional filming equipment and quickly create various video creations in the field of video creation; Create a unique live streaming room and versatile virtual anchor in the live streaming sales scene.
Reading: 30 2024-11-09
360 Intelligent Brain is a large-scale language model developed by 360 Company, with powerful natural language processing and generation capabilities, capable of completing various tasks such as chat interaction, text generation, language understanding, and answering questions. It is an important component of 360's artificial intelligence strategy, aimed at providing users with more efficient and convenient intelligent services.
Reading: 52 2024-11-09
The goal of this project is to promote the development of the open source community for Chinese dialogue big models, with the vision of becoming an LLM Engine that can help everyone. Compared to how to do well in pre training of large language models, BELLE focuses more on how to help everyone obtain their own language model with the best possible instruction expression ability on the basis of open-source pre training of large language models, and reduce the research and application threshold of large language models, especially Chinese large language models. To this end, the BELLE project will continue to open up instruction training data, related models, training code, application scenarios, etc., and will also continuously evaluate the impact of different training data, training algorithms, etc. on model performance. BELLE has been optimized for Chinese, and model tuning only uses data produced by ChatGPT (excluding any other data).
Reading: 54 2024-11-09
The Zhiyuan "Enlightenment" artificial intelligence model has continuously set the record of "China's first+world's largest" and achieved international cutting-edge scientific research breakthroughs. In June 2023, Wudao 3.0 will enter a new stage of comprehensive open source. The Enlightenment Model is an active promoter of the transition from the 'Great Refining Model' to the 'Great Refining Model'.
Reading: 31 2024-11-09
Tsinghua University NLP Laboratory, Face Wall Intelligence, and Zhihu jointly launched the OpenBMB open source multimodal large model series VisCPM. Evaluation shows that VisCPM achieves the best level in Chinese multimodal open source models. VisCPM is an open-source multimodal large model series that supports bilingual multimodal dialogue capability (VisCPM Chat model) and text to image generation capability (VisCPM Paint model). VisCPM is trained on the billion parameter language model CPM Bee (10B), which integrates a visual encoder (Q-Former) and a visual decoder (Diffusion UNet) to support the input and output of visual signals. VisCPM can achieve excellent Chinese multimodal capability through pre training with only English multimodal data and generalization.
Reading: 49 2024-11-09
InternLM is a large-scale multilingual language model developed by Shanghai Artificial Intelligence Laboratory and SenseTime (with equal contributions) in collaboration with The Chinese University of Hong Kong, Fudan University, and Shanghai Jiao Tong University. InternLM is a multilingual base language model with 104B parameters. InternLM undergoes a multi-stage progressive process of pre training on a large corpus with 1.6T tokens, followed by fine-tuning to fit human preferences. We have also developed a training system called Uniscale LLM for efficient training of large-scale language models. The evaluation of multiple benchmarks shows that InternLM has achieved state-of-the-art performance in various aspects such as knowledge understanding, reading comprehension, mathematics, and coding. With such comprehensive capabilities, InternLM achieved excellent results in comprehensive exams including MMLU, AGIEval, C-Eval, and GAOKAO Bench without the need for external tools. In these benchmark tests, InternLM not only outperforms open source models significantly, but also achieves superior performance compared to ChatGPT. In addition, InternLM demonstrates excellent understanding of the Chinese language and culture, making it a suitable foundational model for supporting language applications targeting Chinese.
Reading: 254 2024-11-09
On May 13, 2023, domestic AI company Yunzhisheng announced that it will hold a Yunzhisheng Shanhai Big Model and Achievement Conference on May 24, and will also release products and solutions for smart healthcare, smart IoT, and enterprise services based on the Shanhai Big Model.
Reading: 22 2024-11-09
The Mencius pre training model is based on a large-scale pre training language model developed by the team. It can handle multilingual and multimodal data, while supporting multiple comprehension and generation tasks, and can quickly meet the needs of different fields and application scenarios. Committed to providing a new generation of cognitive intelligence platform for global enterprises based on NLP technology
Reading: 45 2024-11-09
Dongni - Deep thinking AI multimodal search engine, leader in multimodal GPT pre trained large models and human-computer interaction technology.
Reading: 41 2024-11-09
Zidong Taichu is a collaboration between Huawei Ascend Computing and Wuhan Institute of Artificial Intelligence and Institute of Automation, Chinese Academy of Sciences, based on the full stack of Ascend Ascend's software and hardware The Taichu series of models is an important achievement of CAS Automation and Huawei in exploring the path of general artificial intelligence. In September 2021, the world's first 100 billion model of three modes of graphic, text and audio was released, and in September 2022, the industry's largest Chinese multimodal dataset, Zidong, was released Taisu, Zidong The release of Taichu will change the current AI R&D paradigm of single model corresponding to single task, realize multimodal unified semantic expression, and have huge market value in commercial applications such as multimodal content understanding, search, recommendation and question answering, speech recognition and synthesis, human-computer interaction and unmanned driving.
Reading: 42 2024-11-09
! [GitHub - SCIR-HI/Huatuo-Llama-M Med China: Repo for BenTsao [original name: HuaTuo], Llama-7B tuned with Chinese medical knowledge. Bencao (formerly known as HuaTuo) model repository, fine tuned LLaMA model instructions based on Chinese medical knowledge] Bencao [formerly known as HuaTuo]: LLaMA fine-tuning model based on Chinese medical knowledge This project has open-source the LLaMA-7B model that has undergone Chinese medical instruction fine-tuning/instruction fine-tuning. We constructed a Chinese medical instruction dataset using a medical knowledge graph and GPT3.5 API, and based on this, fine tuned LLaMA's instructions to improve its question answering effectiveness in the medical field. Based on the same data, we also trained a medical version of the ChatGLM model: ChatGLM-6B Med In addition, we also attempted to incorporate the conclusions from medical literature as external information into multiple rounds of conversations using the GPT3.5 API, and fine tuned LLaMA instructions based on this. At present. We only open model parameters trained for a single disease of 'liver cancer'. In the future, we plan to release a medical dialogue dataset that incorporates literature conclusions and train models for 16 diseases related to liver, gallbladder, and pancreas.
Reading: 38 2024-11-09
TechGPT is a vertical domain big language model released by the Northeastern University Knowledge Graph Research Group. Currently, a fully fine tuned 7B version has been open sourced on HuggingFace: neukg/TechGPT-7B. TechGPT has mainly strengthened the following three types of tasks: Various information extraction tasks such as relation triplet extraction with "knowledge graph construction" as the core Various intelligent Q&A tasks with "reading comprehension" as the core. Various sequence generation tasks such as keyword generation with "text understanding" as the core. Within these three core capabilities of natural language processing, TechGPT also possesses the ability to process natural language texts in over ten vertical professional fields, including computer science, materials, machinery, metallurgy, finance, and aerospace. At present, TechGPT supports single round and multi round conversations through different prompts and instruction input methods, covering domain term extraction, named entity recognition, relation triplet extraction, text keyword generation, title generation summary, summary generation title, text domain recognition, machine reading comprehension, basic knowledge Q&A, context based knowledge Q&A, advice and consultation Q&A, copy generation, Chinese English translation, and simple code generation, among other natural language understanding and generation capabilities.
Reading: 46 2024-11-09