|
--- |
|
language: |
|
- en |
|
license: mit |
|
library_name: transformers |
|
tags: |
|
- LCARS |
|
- Star-Trek |
|
- 128k-Context |
|
- mistral |
|
- chemistry |
|
- biology |
|
- finance |
|
- legal |
|
- art |
|
- code |
|
- medical |
|
- text-generation-inference |
|
pipeline_tag: text2text-generation |
|
model-index: |
|
- name: LCARS_AI_StarTrek_Computer |
|
results: |
|
- task: |
|
type: text-generation |
|
name: Text Generation |
|
dataset: |
|
name: IFEval (0-Shot) |
|
type: HuggingFaceH4/ifeval |
|
args: |
|
num_few_shot: 0 |
|
metrics: |
|
- type: inst_level_strict_acc and prompt_level_strict_acc |
|
value: 35.83 |
|
name: strict accuracy |
|
source: |
|
url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=LeroyDyer/LCARS_AI_StarTrek_Computer |
|
name: Open LLM Leaderboard |
|
- task: |
|
type: text-generation |
|
name: Text Generation |
|
dataset: |
|
name: BBH (3-Shot) |
|
type: BBH |
|
args: |
|
num_few_shot: 3 |
|
metrics: |
|
- type: acc_norm |
|
value: 21.78 |
|
name: normalized accuracy |
|
source: |
|
url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=LeroyDyer/LCARS_AI_StarTrek_Computer |
|
name: Open LLM Leaderboard |
|
- task: |
|
type: text-generation |
|
name: Text Generation |
|
dataset: |
|
name: MATH Lvl 5 (4-Shot) |
|
type: hendrycks/competition_math |
|
args: |
|
num_few_shot: 4 |
|
metrics: |
|
- type: exact_match |
|
value: 4.08 |
|
name: exact match |
|
source: |
|
url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=LeroyDyer/LCARS_AI_StarTrek_Computer |
|
name: Open LLM Leaderboard |
|
- task: |
|
type: text-generation |
|
name: Text Generation |
|
dataset: |
|
name: GPQA (0-shot) |
|
type: Idavidrein/gpqa |
|
args: |
|
num_few_shot: 0 |
|
metrics: |
|
- type: acc_norm |
|
value: 2.35 |
|
name: acc_norm |
|
source: |
|
url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=LeroyDyer/LCARS_AI_StarTrek_Computer |
|
name: Open LLM Leaderboard |
|
- task: |
|
type: text-generation |
|
name: Text Generation |
|
dataset: |
|
name: MuSR (0-shot) |
|
type: TAUR-Lab/MuSR |
|
args: |
|
num_few_shot: 0 |
|
metrics: |
|
- type: acc_norm |
|
value: 7.44 |
|
name: acc_norm |
|
source: |
|
url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=LeroyDyer/LCARS_AI_StarTrek_Computer |
|
name: Open LLM Leaderboard |
|
- task: |
|
type: text-generation |
|
name: Text Generation |
|
dataset: |
|
name: MMLU-PRO (5-shot) |
|
type: TIGER-Lab/MMLU-Pro |
|
config: main |
|
split: test |
|
args: |
|
num_few_shot: 5 |
|
metrics: |
|
- type: acc |
|
value: 16.2 |
|
name: accuracy |
|
source: |
|
url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=LeroyDyer/LCARS_AI_StarTrek_Computer |
|
name: Open LLM Leaderboard |
|
--- |
|
If anybody has star trek data please send as this starship computer database archive needs it! |
|
|
|
then i can correctly theme this model to be inside its role as a starship computer : |
|
so as well as any space dara ffrom nasa ; i have collected some mufon files which i am still framing the correct prompts for ; for recall as well as interogation : |
|
I shall also be adding a lot of biblical data and historical data ; from sacred texts; so any generated discussions as phylosophers discussing ancient history and how to solve the problems of the past which they encountered ; in thier lifes: using historical and factual data; as well as playig thier roles after generating a biography and character role to the models to play: they should also be amazed by each others acheivements depending on thier periods: |
|
we need multiple role and characters for these discussions: as well as as much historical facts and historys as possible to enhance this models abitlity to dicern ancient aliens truth or false : (so we need astrological, astronomical, as well as sizmological and ecological data for the periods of histroy we know : as well as the unfounded suupositions from youtube subtitles !) another useful source of themed data! |
|
|
|
|
|
This model is a Collection of merged models via various merge methods : Reclaiming Previous models which will be orphened by thier parent models : |
|
THis model is the model of models so it may not Remember some task or Infact remember them all as well as highly perform ! |
|
There were some very bad NSFW Merges from role play to erotica as well as various characters and roles downloaded into the model: |
|
So those models were merged into other models which had been specifically trained for maths or medical data and the coding operations or even translation: |
|
|
|
the models were heavliy dpo trained ; and various newer methodologies installed : the deep mind series is a special series which contains self correction recal, visio spacial ... step by step thinking: |
|
|
|
SO the multi merge often fizes these errors between models as well as training gaps :Hopefully they all took and merged well ! |
|
Performing even unknown and unprogrammed tasks: |
|
# [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard) |
|
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/LeroyDyer__LCARS_AI_StarTrek_Computer-details) |
|
|
|
| Metric |Value| |
|
|-------------------|----:| |
|
|Avg. |14.61| |
|
|IFEval (0-Shot) |35.83| |
|
|BBH (3-Shot) |21.78| |
|
|MATH Lvl 5 (4-Shot)| 4.08| |
|
|GPQA (0-shot) | 2.35| |
|
|MuSR (0-shot) | 7.44| |
|
|MMLU-PRO (5-shot) |16.20| |
|
|
|
|