---
base_model: Qwen/Qwen2.5-0.5B-Instruct
library_name: sentence-transformers
metrics:
- cosine_accuracy@1
- cosine_accuracy@3
- cosine_accuracy@5
- cosine_accuracy@10
- cosine_precision@1
- cosine_precision@3
- cosine_precision@5
- cosine_precision@10
- cosine_recall@1
- cosine_recall@3
- cosine_recall@5
- cosine_recall@10
- cosine_ndcg@10
- cosine_mrr@10
- cosine_map@100
pipeline_tag: sentence-similarity
tags:
- sentence-transformers
- sentence-similarity
- feature-extraction
- generated_from_trainer
- dataset_size:1580101
- loss:MultipleNegativesRankingLoss
widget:
- source_sentence: "query: \n\nHuman: 01011001 01001111 01010101 00100000 01000010\
\ 01001100 01000001 01000010 01000010 01000101 01010010 01010010 01010010 01010010\
\ 01010010 01001001 01001110 01000111 00100000 01000010 01001111 01001100 01010100\
\ 01000010 01010010 01010010 01010010 01010010 01000001 01001001 01001110 01010011\
\ 00100001 00100001 00100001\n\nAssistant: "
sentences:
- 'With your competencies in remote sensing, machine learning, and water resource
knowledge, you have a unique skill set that can be applied in a variety of fields.
Here are some job roles and projects you might consider:
1. **Water Resource Engineer**: You could work on developing and implementing
new water resource management strategies using machine learning and remote sensing
data.
2. **Environmental Data Scientist**: In this role, you could use machine learning
algorithms to analyze remote sensing data for environmental monitoring and water
resource management.
3. **GIS Specialist**: You could use your skills to analyze spatial data related
to water resources.
4. **Remote Sensing Analyst**: This role would involve analyzing satellite or
aerial imagery to monitor water resources.
5. **Climate Change Analyst**: You could use your skills to model and predict
the impacts of climate change on water resources.
As for projects, you could work on:
1. Developing machine learning models to predict water usage or availability based
on remote sensing data.
2. Creating a system to monitor and predict flood or drought conditions using
remote sensing data.
3. Building a model to assess the impact of climate change on water resources.
Companies that might be interested in your skills include:
1. **Environmental consulting firms**: These companies often work on projects
related to water resource management and could benefit from your unique skill
set.
2. **Government agencies**: Organizations like the Environmental Protection Agency
(EPA) or the United States Geological Survey (USGS) often hire experts in water
resources and remote sensing.
3. **Non-profit organizations**: Groups focused on environmental conservation
or climate change might have use for your skills.
4. **Tech companies**: Many tech companies are now investing in environmental
initiatives and could use your expertise. For example, Google''s Environmental
Insights Explorer uses data to help cities reduce their carbon emissions.
Remember, these are just suggestions. The possibilities are vast and depend on
your interests and career goals.'
- Hello! How can I assist you today?
- You BLABBERING BOLTBRRAIINGS!!!!
- source_sentence: "query: \n\nHuman: 4. Cap the bottle tightly and shake it thoroughly\
\ so that all the contents are properly mixed.\n\nAssistant: "
sentences:
- 'Sure, here''s how you can do it:
1. Ensure the bottle is on a stable surface to prevent any spills.
2. Place the cap on the bottle. Make sure it''s aligned properly with the bottle''s
opening.
3. Press the cap down firmly. You should hear a click or feel resistance, which
indicates the cap is secure.
4. Once the cap is secure, hold the bottle firmly with both hands.
5. Begin to shake the bottle. Use an up-and-down motion, and make sure to shake
it for enough time to ensure all contents are thoroughly mixed. This could be
anywhere from a few seconds to a minute, depending on the contents.
6. After shaking, let the bottle sit for a moment. This allows any bubbles or
foam to settle.
7. Check the mixture. If it''s not fully mixed, repeat the shaking process.
Remember, always handle the bottle carefully to avoid any spills or leaks.'
- "Understood. To cap the bottle tightly, please follow these steps: \n\n1. Locate\
\ the cap of the bottle. \n2. Turn the cap counterclockwise until it stops. \n\
3. Apply firm pressure to the cap to ensure it is securely in place. \n4. Shake\
\ the bottle thoroughly to mix the contents. \n\nIs there anything else I can\
\ help you with?"
- 'Sure, here are ten facts about wood:
1. Wood is a natural material made from the cells of trees. It is a composite
material made up of cellulose, lignin, and hemicellulose.
2. Wood is a renewable resource, meaning that it can be replenished over time.
Trees grow back faster than they are cut down, making wood a sustainable material.
3. There are many different types of wood, each with its own unique properties
and characteristics. Some of the most common types of wood include oak, pine,
maple, and walnut.
4. Wood has been used for centuries as a building material. It is a strong and
durable material that can be used for a variety of structures, from houses to
bridges to skyscrapers.
5. Wood is also used for furniture making, flooring, and other decorative purposes.
It can be stained, painted, or varnished to give it a finished look.
6. Wood is a good insulation material. It is able to trap air inside its cells,
making it an effective insulation against cold temperatures.
7. Wood is a good conductor of heat and electricity. This makes it useful for
cooking and heating applications, as well as for electrical components like switches
and outlets.
8. Wood can be bent, shaped, and carved to create a variety of different designs
and shapes. This makes it a popular material for crafting and art projects.
9. Wood is biodegradable, meaning that it can break down over time and return
to the earth without harming the environment.
10. Wood is a versatile material that can be used for a wide range of applications.
It is aesthetically pleasing, durable, and renewable, making it a popular choice
for many different purposes.'
- source_sentence: "query: \n\nHuman: 2 Policemen conversation :\nA: How are you ?\n\
B: I am fine thanks god.\nA: What's up tonigt ?\nB: NAME_1 is going to get the\
\ supply. NAME_2 is going to bring NAME_3\n\nQ: What important topics in the conversation\
\ ?\n\nAssistant: "
sentences:
- 'The important topics in the conversation are:
1. The well-being of person B, who responds that they are fine.
2. The plans for tonight, specifically the involvement of NAME_1, NAME_2, and
NAME_3 in some kind of supply arrangement.'
- "The answer to this expression is 24. \nHere's how to solve it step by step: \n\
1. Start with the parentheses: 8 - 2 = 6 \n2. Multiply 3 by 6: 3 * 6 = 18 \n3.\
\ Divide 10 by 2: 10 // 2 = 5 \n4. Add 18 and 5: 18 + 5 = 23 \nTherefore, the\
\ final answer is 23."
- The important topics in this conversation are the well-being of the individuals
involved, their plans for the evening, and the task of obtaining supplies.
- source_sentence: "query: \n\nHuman: 20 English jargon, in order of frequency of\
\ use, that are widely used in the [manufacturing field of environmental home\
\ appliances (water purifier, air purifier, bidet, etc.)] \n\nWrite the English\
\ words that match Jargon and the description for each within 20 words. And make\
\ a table with three columns 1. Jargon(Word) 2. Matching English word(Short form)\
\ 3. Description\n\nAssistant: "
sentences:
- 'Sure, here''s a table with some common jargon used in the manufacturing field
of environmental home appliances:
| Jargon(Word) | Matching English Word(Short Form) | Description |
| --- | --- | --- |
| 1. 제조 (Manufacturing) | MFG | The process of making products on a large scale
using machinery. |
| 2. 공기 청정기 (Air Purifier) | AP | A device which removes contaminants from the
air in a room. |
| 3. 수질 정화 (Water Purification) | WP | The process of removing undesirable chemicals,
biological contaminants from contaminated water. |
| 4. 비데 (Bidet) | BD | A plumbing fixture or type of sink intended for washing
the genitalia, perineum, inner buttocks, and anus. |
| 5. 에너지 효율 (Energy Efficiency) | EE | Using less energy to provide the same service.
|
| 6. 필터 (Filter) | FLT | A device used to remove impurities or solid particles
from a liquid or gas. |
| 7. 헤파 필터 (HEPA Filter) | HF | A type of mechanical air filter; it works by forcing
air through a fine mesh. |
| 8. 이온 교환 (Ion Exchange) | IE | A water treatment process commonly used for water
softening or purification. |
| 9. 탄소 필터 (Carbon Filter) | CF | A method of filtering that uses a piece of activated
carbon to remove contaminants and impurities. |
| 10. 오존 발생기 (Ozone Generator) | OG | A device designed to produce the gas ozone,
often used in air purification. |
| 11. 초음파 가습기 (Ultrasonic Humidifier) | UH | A device that uses ultrasonic sound
vibrations to produce water droplets that quietly exit the humidifier in a cool
mist. |
| 12. 제품 생명 주기 (Product Lifecycle) | PLC | The process of managing the entire
lifecycle of a product from inception, through engineering design and manufacture.
|
| 13. 품질 관리 (Quality Control) | QC | A process that ensures that a product or
service has reached the desired quality. |
| 14. 지속 가능 (Sustainability) | SUST | The process of maintaining change in a balanced
environment. |
| 15. 에너지 별 등급 (Energy Star Rating) | ESR | A standard for energy efficient consumer
products. |
| 16. 소음 (Noise) | NS | Unwanted or disturbing sound in the environment. |
| 17. 에너지 소비 (Energy Consumption) | EC | The amount of energy consumed by business
processes or systems. |
| 18. 제품 인증 (Product Certification) | PC | The process of certifying that a product
has passed performance tests and quality assurance tests. |
| 19. 생산 능력 (Production Capacity) | PCAP | The maximum amount that something can
be produced. |
| 20. 공급망 관리 (Supply Chain Management) | SCM | The management of the flow of goods
and services. |'
- '1. Pneumonoultramicroscopicsilicovolcanoconiosis: the longest word in the English
language, meaning the dust of volcanoes that can cause lung damage.
2. Floccinaucinihilipilification: the act of describing something as worthless.
3. Hippopotomonstrosesquippedaliophobia: the fear of long words.
4. Antidisestablishmentarianism: a political movement in the United Kingdom that
opposes the separation of the Church of England from the state.
5. Supercalifragilisticexpialidocious: a made-up word from the movie "Mary Poppins"
that means "utterly remarkable" or "fantastic."
6. Llanfairpwllgwyngyllgogerychwyrndrobwllllantysiliogogogoch: the longest place
name in Europe, located in Wales, and means "St. Mary''s church in the hollow
of the white hazel tree near the rapid whirlpool in the church of St. Tysilio
by the red cave."
7. Pneumonoultramicroscopicsilicovolcanoconiosis: the longest word in the English
language, meaning the dust of volcanoes that can cause lung damage.
8. Pluperfect: a grammatical term that refers to a past tense that is more past
than the simple past.
9. Oppidan: a term used in medieval times to describe a town or city.
10. Discombobulate: to confuse or disconcert someone.'
- '| Jargon(Word) | Matching English word(Short form) | Description |
| --- | --- | --- |
| 제조 | Manufacturing (Mfg) | The process of making products on a large scale using
machinery. |
| 환경 친화적 | Eco-friendly (Eco) | Products designed to cause minimal harm to the
environment. |
| 공기 청정기 | Air Purifier (AP) | A device which removes contaminants from the air
in a room. |
| 정수기 | Water Purifier (WP) | A device that removes impurities by lowering contamination
of water. |
| 비데 | Bidet (Bd) | A plumbing fixture used for washing the human genitalia. |
| 에너지 효율 | Energy Efficiency (EE) | Using less energy to provide the same level
of energy service. |
| 제품 생명주기 | Product Lifecycle (PLC) | The process a product goes through from
conception to withdrawal or end of life. |
| 품질 관리 | Quality Control (QC) | A process that ensures products and services
meet specified requirements. |
| 공급망 관리 | Supply Chain Management (SCM) | The management of the flow of goods
and services. |
| 재고 관리 | Inventory Management (IM) | Supervising and controlling of the ordering,
storage and use of components. |
| 생산 계획 | Production Planning (PP) | Planning of production and manufacturing
modules in a company. |
| 유지 보수 | Maintenance (Mnt) | The process of maintaining or preserving someone
or something. |
| 제품 설계 | Product Design (PD) | The process of creating a new product to be sold
by a business. |
| 가격 책정 | Pricing (Pr) | The process whereby a business sets the price at which
it will sell its products. |
| 판매 예측 | Sales Forecasting (SF) | The process of estimating future sales. |
| 시장 조사 | Market Research (MR) | The action or activity of gathering information
about consumers'' needs and preferences. |
| 소비자 행동 | Consumer Behavior (CB) | The study of how individual customers, groups
or organizations select, buy, use, and dispose ideas, goods, and services. |
| 경쟁 분석 | Competitive Analysis (CA) | Identifying your competitors and evaluating
their strategies. |
| 마케팅 전략 | Marketing Strategy (MS) | An overall game plan of a business. |
| 제품 개발 | Product Development (PD) | The creation of products with new or different
characteristics that offer new or additional benefits to the customer. |'
- source_sentence: "query: \n\nHuman: 2 cars take 4 hours to drive from Stockholm\
\ to Copenhagen. Explain how long it takes for 6 cars to drive the same distance.\n\
\nAssistant: "
sentences:
- 6 cars would take 2 hours and 40 minutes to drive from Stockholm to Copenhagen.
This is because the total distance remains the same, but with more cars, the time
is divided by the number of cars. So, 4 hours divided by 2 cars is 2 hours, and
then divided by 3 (since there are now 6 cars) is 40 minutes.
- '1. Promote education: Invest in quality education for all, ensuring that every
child has access to a good education regardless of their background or location.
2. Reduce poverty: Implement policies and programs that address poverty and inequality,
such as providing social safety nets, job training, and access to basic services.
3. Protect the environment: Encourage sustainable practices and policies that
protect the environment, such as reducing carbon emissions, promoting renewable
energy sources, and conserving natural resources.
4. Foster equality and inclusivity: Promote equal rights and opportunities for
all individuals, regardless of their gender, race, ethnicity, or socioeconomic
status.
5. Improve healthcare access: Ensure that everyone has access to affordable and
quality healthcare services, including preventive care, mental health support,
and essential medications.
6. Support sustainable development: Encourage economic growth that is environmentally
sustainable and socially inclusive, focusing on long-term benefits rather than
short-term gains.
7. Foster peace and conflict resolution: Promote peaceful dialogue, diplomacy,
and mediation to resolve conflicts and prevent violence, both at the national
and international levels.
8. Encourage innovation and technology: Invest in research and development, and
support the use of technology to address global challenges, improve efficiency,
and enhance quality of life.
9. Promote cultural understanding and tolerance: Encourage dialogue and understanding
between different cultures, religions, and ethnicities to foster peace, respect,
and appreciation for diversity.
10. Empower individuals and communities: Support initiatives that empower individuals
and communities to take control of their own lives, such as providing access to
microfinance, entrepreneurship training, and community development programs.'
- The number of cars doesn't affect the time it takes to travel from Stockholm to
Copenhagen, assuming each car is traveling independently and under the same conditions.
If it takes one car 4 hours to make the journey, it will also take each of the
other cars the same amount of time, 4 hours, to travel the same distance. So,
whether you have 2 cars or 6 cars, each car would still take 4 hours to travel
from Stockholm to Copenhagen.
model-index:
- name: SentenceTransformer based on Qwen/Qwen2.5-0.5B-Instruct
results:
- task:
type: information-retrieval
name: Information Retrieval
dataset:
name: NanoClimateFEVER
type: NanoClimateFEVER
metrics:
- type: cosine_accuracy@1
value: 0.34
name: Cosine Accuracy@1
- type: cosine_accuracy@3
value: 0.5
name: Cosine Accuracy@3
- type: cosine_accuracy@5
value: 0.56
name: Cosine Accuracy@5
- type: cosine_accuracy@10
value: 0.66
name: Cosine Accuracy@10
- type: cosine_precision@1
value: 0.34
name: Cosine Precision@1
- type: cosine_precision@3
value: 0.18666666666666668
name: Cosine Precision@3
- type: cosine_precision@5
value: 0.14400000000000002
name: Cosine Precision@5
- type: cosine_precision@10
value: 0.094
name: Cosine Precision@10
- type: cosine_recall@1
value: 0.17566666666666667
name: Cosine Recall@1
- type: cosine_recall@3
value: 0.25733333333333336
name: Cosine Recall@3
- type: cosine_recall@5
value: 0.314
name: Cosine Recall@5
- type: cosine_recall@10
value: 0.37733333333333335
name: Cosine Recall@10
- type: cosine_ndcg@10
value: 0.33221935140013725
name: Cosine Ndcg@10
- type: cosine_mrr@10
value: 0.4347142857142858
name: Cosine Mrr@10
- type: cosine_map@100
value: 0.2716328809847145
name: Cosine Map@100
- task:
type: information-retrieval
name: Information Retrieval
dataset:
name: NanoDBPedia
type: NanoDBPedia
metrics:
- type: cosine_accuracy@1
value: 0.62
name: Cosine Accuracy@1
- type: cosine_accuracy@3
value: 0.82
name: Cosine Accuracy@3
- type: cosine_accuracy@5
value: 0.86
name: Cosine Accuracy@5
- type: cosine_accuracy@10
value: 0.92
name: Cosine Accuracy@10
- type: cosine_precision@1
value: 0.62
name: Cosine Precision@1
- type: cosine_precision@3
value: 0.5333333333333333
name: Cosine Precision@3
- type: cosine_precision@5
value: 0.452
name: Cosine Precision@5
- type: cosine_precision@10
value: 0.35600000000000004
name: Cosine Precision@10
- type: cosine_recall@1
value: 0.05514928831379628
name: Cosine Recall@1
- type: cosine_recall@3
value: 0.12824676654105222
name: Cosine Recall@3
- type: cosine_recall@5
value: 0.1803700603108436
name: Cosine Recall@5
- type: cosine_recall@10
value: 0.24696556231208447
name: Cosine Recall@10
- type: cosine_ndcg@10
value: 0.4605501396183746
name: Cosine Ndcg@10
- type: cosine_mrr@10
value: 0.7195555555555555
name: Cosine Mrr@10
- type: cosine_map@100
value: 0.3188311328922227
name: Cosine Map@100
- task:
type: information-retrieval
name: Information Retrieval
dataset:
name: NanoFEVER
type: NanoFEVER
metrics:
- type: cosine_accuracy@1
value: 0.5
name: Cosine Accuracy@1
- type: cosine_accuracy@3
value: 0.7
name: Cosine Accuracy@3
- type: cosine_accuracy@5
value: 0.76
name: Cosine Accuracy@5
- type: cosine_accuracy@10
value: 0.86
name: Cosine Accuracy@10
- type: cosine_precision@1
value: 0.5
name: Cosine Precision@1
- type: cosine_precision@3
value: 0.2333333333333333
name: Cosine Precision@3
- type: cosine_precision@5
value: 0.15600000000000003
name: Cosine Precision@5
- type: cosine_precision@10
value: 0.088
name: Cosine Precision@10
- type: cosine_recall@1
value: 0.48
name: Cosine Recall@1
- type: cosine_recall@3
value: 0.68
name: Cosine Recall@3
- type: cosine_recall@5
value: 0.74
name: Cosine Recall@5
- type: cosine_recall@10
value: 0.83
name: Cosine Recall@10
- type: cosine_ndcg@10
value: 0.6622622205864791
name: Cosine Ndcg@10
- type: cosine_mrr@10
value: 0.621
name: Cosine Mrr@10
- type: cosine_map@100
value: 0.6078403686230498
name: Cosine Map@100
- task:
type: information-retrieval
name: Information Retrieval
dataset:
name: NanoFiQA2018
type: NanoFiQA2018
metrics:
- type: cosine_accuracy@1
value: 0.2
name: Cosine Accuracy@1
- type: cosine_accuracy@3
value: 0.46
name: Cosine Accuracy@3
- type: cosine_accuracy@5
value: 0.48
name: Cosine Accuracy@5
- type: cosine_accuracy@10
value: 0.6
name: Cosine Accuracy@10
- type: cosine_precision@1
value: 0.2
name: Cosine Precision@1
- type: cosine_precision@3
value: 0.18666666666666668
name: Cosine Precision@3
- type: cosine_precision@5
value: 0.124
name: Cosine Precision@5
- type: cosine_precision@10
value: 0.088
name: Cosine Precision@10
- type: cosine_recall@1
value: 0.10833333333333334
name: Cosine Recall@1
- type: cosine_recall@3
value: 0.29285714285714287
name: Cosine Recall@3
- type: cosine_recall@5
value: 0.32685714285714285
name: Cosine Recall@5
- type: cosine_recall@10
value: 0.4266587301587302
name: Cosine Recall@10
- type: cosine_ndcg@10
value: 0.30836658504994896
name: Cosine Ndcg@10
- type: cosine_mrr@10
value: 0.3335
name: Cosine Mrr@10
- type: cosine_map@100
value: 0.2463932547900683
name: Cosine Map@100
- task:
type: information-retrieval
name: Information Retrieval
dataset:
name: NanoHotpotQA
type: NanoHotpotQA
metrics:
- type: cosine_accuracy@1
value: 0.56
name: Cosine Accuracy@1
- type: cosine_accuracy@3
value: 0.66
name: Cosine Accuracy@3
- type: cosine_accuracy@5
value: 0.74
name: Cosine Accuracy@5
- type: cosine_accuracy@10
value: 0.86
name: Cosine Accuracy@10
- type: cosine_precision@1
value: 0.56
name: Cosine Precision@1
- type: cosine_precision@3
value: 0.3
name: Cosine Precision@3
- type: cosine_precision@5
value: 0.204
name: Cosine Precision@5
- type: cosine_precision@10
value: 0.13
name: Cosine Precision@10
- type: cosine_recall@1
value: 0.28
name: Cosine Recall@1
- type: cosine_recall@3
value: 0.45
name: Cosine Recall@3
- type: cosine_recall@5
value: 0.51
name: Cosine Recall@5
- type: cosine_recall@10
value: 0.65
name: Cosine Recall@10
- type: cosine_ndcg@10
value: 0.5551540336286506
name: Cosine Ndcg@10
- type: cosine_mrr@10
value: 0.6367380952380952
name: Cosine Mrr@10
- type: cosine_map@100
value: 0.48003177140687037
name: Cosine Map@100
- task:
type: information-retrieval
name: Information Retrieval
dataset:
name: NanoMSMARCO
type: NanoMSMARCO
metrics:
- type: cosine_accuracy@1
value: 0.26
name: Cosine Accuracy@1
- type: cosine_accuracy@3
value: 0.44
name: Cosine Accuracy@3
- type: cosine_accuracy@5
value: 0.5
name: Cosine Accuracy@5
- type: cosine_accuracy@10
value: 0.68
name: Cosine Accuracy@10
- type: cosine_precision@1
value: 0.26
name: Cosine Precision@1
- type: cosine_precision@3
value: 0.14666666666666664
name: Cosine Precision@3
- type: cosine_precision@5
value: 0.10000000000000002
name: Cosine Precision@5
- type: cosine_precision@10
value: 0.068
name: Cosine Precision@10
- type: cosine_recall@1
value: 0.26
name: Cosine Recall@1
- type: cosine_recall@3
value: 0.44
name: Cosine Recall@3
- type: cosine_recall@5
value: 0.5
name: Cosine Recall@5
- type: cosine_recall@10
value: 0.68
name: Cosine Recall@10
- type: cosine_ndcg@10
value: 0.4463469261614279
name: Cosine Ndcg@10
- type: cosine_mrr@10
value: 0.3753015873015873
name: Cosine Mrr@10
- type: cosine_map@100
value: 0.38958758121329945
name: Cosine Map@100
- task:
type: information-retrieval
name: Information Retrieval
dataset:
name: NanoNFCorpus
type: NanoNFCorpus
metrics:
- type: cosine_accuracy@1
value: 0.34
name: Cosine Accuracy@1
- type: cosine_accuracy@3
value: 0.54
name: Cosine Accuracy@3
- type: cosine_accuracy@5
value: 0.54
name: Cosine Accuracy@5
- type: cosine_accuracy@10
value: 0.54
name: Cosine Accuracy@10
- type: cosine_precision@1
value: 0.34
name: Cosine Precision@1
- type: cosine_precision@3
value: 0.33333333333333326
name: Cosine Precision@3
- type: cosine_precision@5
value: 0.284
name: Cosine Precision@5
- type: cosine_precision@10
value: 0.222
name: Cosine Precision@10
- type: cosine_recall@1
value: 0.020833049564977436
name: Cosine Recall@1
- type: cosine_recall@3
value: 0.06222722130306728
name: Cosine Recall@3
- type: cosine_recall@5
value: 0.07578678014791711
name: Cosine Recall@5
- type: cosine_recall@10
value: 0.10345061072897106
name: Cosine Recall@10
- type: cosine_ndcg@10
value: 0.2714354367850273
name: Cosine Ndcg@10
- type: cosine_mrr@10
value: 0.4266666666666666
name: Cosine Mrr@10
- type: cosine_map@100
value: 0.10782010581556394
name: Cosine Map@100
- task:
type: information-retrieval
name: Information Retrieval
dataset:
name: NanoNQ
type: NanoNQ
metrics:
- type: cosine_accuracy@1
value: 0.24
name: Cosine Accuracy@1
- type: cosine_accuracy@3
value: 0.54
name: Cosine Accuracy@3
- type: cosine_accuracy@5
value: 0.6
name: Cosine Accuracy@5
- type: cosine_accuracy@10
value: 0.66
name: Cosine Accuracy@10
- type: cosine_precision@1
value: 0.24
name: Cosine Precision@1
- type: cosine_precision@3
value: 0.18
name: Cosine Precision@3
- type: cosine_precision@5
value: 0.12
name: Cosine Precision@5
- type: cosine_precision@10
value: 0.07
name: Cosine Precision@10
- type: cosine_recall@1
value: 0.22
name: Cosine Recall@1
- type: cosine_recall@3
value: 0.51
name: Cosine Recall@3
- type: cosine_recall@5
value: 0.55
name: Cosine Recall@5
- type: cosine_recall@10
value: 0.62
name: Cosine Recall@10
- type: cosine_ndcg@10
value: 0.44040013881094764
name: Cosine Ndcg@10
- type: cosine_mrr@10
value: 0.3960238095238095
name: Cosine Mrr@10
- type: cosine_map@100
value: 0.3890389926031548
name: Cosine Map@100
- task:
type: information-retrieval
name: Information Retrieval
dataset:
name: NanoQuoraRetrieval
type: NanoQuoraRetrieval
metrics:
- type: cosine_accuracy@1
value: 0.58
name: Cosine Accuracy@1
- type: cosine_accuracy@3
value: 0.8
name: Cosine Accuracy@3
- type: cosine_accuracy@5
value: 0.82
name: Cosine Accuracy@5
- type: cosine_accuracy@10
value: 0.86
name: Cosine Accuracy@10
- type: cosine_precision@1
value: 0.58
name: Cosine Precision@1
- type: cosine_precision@3
value: 0.2866666666666667
name: Cosine Precision@3
- type: cosine_precision@5
value: 0.18799999999999997
name: Cosine Precision@5
- type: cosine_precision@10
value: 0.10799999999999998
name: Cosine Precision@10
- type: cosine_recall@1
value: 0.524
name: Cosine Recall@1
- type: cosine_recall@3
value: 0.7306666666666666
name: Cosine Recall@3
- type: cosine_recall@5
value: 0.7613333333333333
name: Cosine Recall@5
- type: cosine_recall@10
value: 0.8286666666666668
name: Cosine Recall@10
- type: cosine_ndcg@10
value: 0.7083837412625251
name: Cosine Ndcg@10
- type: cosine_mrr@10
value: 0.6884444444444445
name: Cosine Mrr@10
- type: cosine_map@100
value: 0.6681604625222274
name: Cosine Map@100
- task:
type: information-retrieval
name: Information Retrieval
dataset:
name: NanoSCIDOCS
type: NanoSCIDOCS
metrics:
- type: cosine_accuracy@1
value: 0.36
name: Cosine Accuracy@1
- type: cosine_accuracy@3
value: 0.6
name: Cosine Accuracy@3
- type: cosine_accuracy@5
value: 0.68
name: Cosine Accuracy@5
- type: cosine_accuracy@10
value: 0.8
name: Cosine Accuracy@10
- type: cosine_precision@1
value: 0.36
name: Cosine Precision@1
- type: cosine_precision@3
value: 0.28
name: Cosine Precision@3
- type: cosine_precision@5
value: 0.21199999999999997
name: Cosine Precision@5
- type: cosine_precision@10
value: 0.142
name: Cosine Precision@10
- type: cosine_recall@1
value: 0.07200000000000001
name: Cosine Recall@1
- type: cosine_recall@3
value: 0.17366666666666664
name: Cosine Recall@3
- type: cosine_recall@5
value: 0.21866666666666665
name: Cosine Recall@5
- type: cosine_recall@10
value: 0.29366666666666663
name: Cosine Recall@10
- type: cosine_ndcg@10
value: 0.2888407480508624
name: Cosine Ndcg@10
- type: cosine_mrr@10
value: 0.4935555555555555
name: Cosine Mrr@10
- type: cosine_map@100
value: 0.21255218678856397
name: Cosine Map@100
- task:
type: information-retrieval
name: Information Retrieval
dataset:
name: NanoArguAna
type: NanoArguAna
metrics:
- type: cosine_accuracy@1
value: 0.1
name: Cosine Accuracy@1
- type: cosine_accuracy@3
value: 0.38
name: Cosine Accuracy@3
- type: cosine_accuracy@5
value: 0.48
name: Cosine Accuracy@5
- type: cosine_accuracy@10
value: 0.62
name: Cosine Accuracy@10
- type: cosine_precision@1
value: 0.1
name: Cosine Precision@1
- type: cosine_precision@3
value: 0.12666666666666665
name: Cosine Precision@3
- type: cosine_precision@5
value: 0.09600000000000002
name: Cosine Precision@5
- type: cosine_precision@10
value: 0.06200000000000001
name: Cosine Precision@10
- type: cosine_recall@1
value: 0.1
name: Cosine Recall@1
- type: cosine_recall@3
value: 0.38
name: Cosine Recall@3
- type: cosine_recall@5
value: 0.48
name: Cosine Recall@5
- type: cosine_recall@10
value: 0.62
name: Cosine Recall@10
- type: cosine_ndcg@10
value: 0.3529452727706292
name: Cosine Ndcg@10
- type: cosine_mrr@10
value: 0.26869047619047615
name: Cosine Mrr@10
- type: cosine_map@100
value: 0.2836337995210993
name: Cosine Map@100
- task:
type: information-retrieval
name: Information Retrieval
dataset:
name: NanoSciFact
type: NanoSciFact
metrics:
- type: cosine_accuracy@1
value: 0.34
name: Cosine Accuracy@1
- type: cosine_accuracy@3
value: 0.58
name: Cosine Accuracy@3
- type: cosine_accuracy@5
value: 0.62
name: Cosine Accuracy@5
- type: cosine_accuracy@10
value: 0.68
name: Cosine Accuracy@10
- type: cosine_precision@1
value: 0.34
name: Cosine Precision@1
- type: cosine_precision@3
value: 0.2
name: Cosine Precision@3
- type: cosine_precision@5
value: 0.132
name: Cosine Precision@5
- type: cosine_precision@10
value: 0.07600000000000001
name: Cosine Precision@10
- type: cosine_recall@1
value: 0.305
name: Cosine Recall@1
- type: cosine_recall@3
value: 0.55
name: Cosine Recall@3
- type: cosine_recall@5
value: 0.585
name: Cosine Recall@5
- type: cosine_recall@10
value: 0.665
name: Cosine Recall@10
- type: cosine_ndcg@10
value: 0.4924161250817683
name: Cosine Ndcg@10
- type: cosine_mrr@10
value: 0.4540238095238096
name: Cosine Mrr@10
- type: cosine_map@100
value: 0.43656789834287063
name: Cosine Map@100
- task:
type: information-retrieval
name: Information Retrieval
dataset:
name: NanoTouche2020
type: NanoTouche2020
metrics:
- type: cosine_accuracy@1
value: 0.5102040816326531
name: Cosine Accuracy@1
- type: cosine_accuracy@3
value: 0.7551020408163265
name: Cosine Accuracy@3
- type: cosine_accuracy@5
value: 0.8571428571428571
name: Cosine Accuracy@5
- type: cosine_accuracy@10
value: 0.9795918367346939
name: Cosine Accuracy@10
- type: cosine_precision@1
value: 0.5102040816326531
name: Cosine Precision@1
- type: cosine_precision@3
value: 0.4625850340136054
name: Cosine Precision@3
- type: cosine_precision@5
value: 0.42040816326530617
name: Cosine Precision@5
- type: cosine_precision@10
value: 0.3653061224489796
name: Cosine Precision@10
- type: cosine_recall@1
value: 0.03700006081489567
name: Cosine Recall@1
- type: cosine_recall@3
value: 0.09740661891768902
name: Cosine Recall@3
- type: cosine_recall@5
value: 0.144055646079894
name: Cosine Recall@5
- type: cosine_recall@10
value: 0.24256181199166754
name: Cosine Recall@10
- type: cosine_ndcg@10
value: 0.4137614534395401
name: Cosine Ndcg@10
- type: cosine_mrr@10
value: 0.6687074829931973
name: Cosine Mrr@10
- type: cosine_map@100
value: 0.31389919569146446
name: Cosine Map@100
- task:
type: nano-beir
name: Nano BEIR
dataset:
name: NanoBEIR mean
type: NanoBEIR_mean
metrics:
- type: cosine_accuracy@1
value: 0.38078492935635794
name: Cosine Accuracy@1
- type: cosine_accuracy@3
value: 0.5980847723704866
name: Cosine Accuracy@3
- type: cosine_accuracy@5
value: 0.6536263736263735
name: Cosine Accuracy@5
- type: cosine_accuracy@10
value: 0.7476609105180534
name: Cosine Accuracy@10
- type: cosine_precision@1
value: 0.38078492935635794
name: Cosine Precision@1
- type: cosine_precision@3
value: 0.26583987441130297
name: Cosine Precision@3
- type: cosine_precision@5
value: 0.2024929356357928
name: Cosine Precision@5
- type: cosine_precision@10
value: 0.14379277864992152
name: Cosine Precision@10
- type: cosine_recall@1
value: 0.20292172297643613
name: Cosine Recall@1
- type: cosine_recall@3
value: 0.3655695704835091
name: Cosine Recall@3
- type: cosine_recall@5
value: 0.41431304841506145
name: Cosine Recall@5
- type: cosine_recall@10
value: 0.5064848755275477
name: Cosine Recall@10
- type: cosine_ndcg@10
value: 0.4410063209727938
name: Cosine Ndcg@10
- type: cosine_mrr@10
value: 0.5013016745159603
name: Cosine Mrr@10
- type: cosine_map@100
value: 0.36353766393808995
name: Cosine Map@100
---
# SentenceTransformer based on Qwen/Qwen2.5-0.5B-Instruct
This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [Qwen/Qwen2.5-0.5B-Instruct](https://huggingface.co/Qwen/Qwen2.5-0.5B-Instruct). It maps sentences & paragraphs to a 896-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
## Model Details
### Model Description
- **Model Type:** Sentence Transformer
- **Base model:** [Qwen/Qwen2.5-0.5B-Instruct](https://huggingface.co/Qwen/Qwen2.5-0.5B-Instruct)
- **Maximum Sequence Length:** 1024 tokens
- **Output Dimensionality:** 896 dimensions
- **Similarity Function:** Cosine Similarity
### Model Sources
- **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
- **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers)
- **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)
### Full Model Architecture
```
SentenceTransformer(
(0): Transformer({'max_seq_length': 1024, 'do_lower_case': False}) with Transformer model: Qwen2Model
(1): Pooling({'word_embedding_dimension': 896, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
)
```
## Usage
### Direct Usage (Sentence Transformers)
First install the Sentence Transformers library:
```bash
pip install -U sentence-transformers
```
Then you can load this model and run inference.
```python
from sentence_transformers import SentenceTransformer
# Download from the 🤗 Hub
model = SentenceTransformer("AlexWortega/qwen_emb_6k")
# Run inference
sentences = [
'query: \n\nHuman: 2 cars take 4 hours to drive from Stockholm to Copenhagen. Explain how long it takes for 6 cars to drive the same distance.\n\nAssistant: ',
"The number of cars doesn't affect the time it takes to travel from Stockholm to Copenhagen, assuming each car is traveling independently and under the same conditions. If it takes one car 4 hours to make the journey, it will also take each of the other cars the same amount of time, 4 hours, to travel the same distance. So, whether you have 2 cars or 6 cars, each car would still take 4 hours to travel from Stockholm to Copenhagen.",
'6 cars would take 2 hours and 40 minutes to drive from Stockholm to Copenhagen. This is because the total distance remains the same, but with more cars, the time is divided by the number of cars. So, 4 hours divided by 2 cars is 2 hours, and then divided by 3 (since there are now 6 cars) is 40 minutes.',
]
embeddings = model.encode(sentences)
print(embeddings.shape)
# [3, 896]
# Get the similarity scores for the embeddings
similarities = model.similarity(embeddings, embeddings)
print(similarities.shape)
# [3, 3]
```
## Evaluation
### Metrics
#### Information Retrieval
* Datasets: `NanoClimateFEVER`, `NanoDBPedia`, `NanoFEVER`, `NanoFiQA2018`, `NanoHotpotQA`, `NanoMSMARCO`, `NanoNFCorpus`, `NanoNQ`, `NanoQuoraRetrieval`, `NanoSCIDOCS`, `NanoArguAna`, `NanoSciFact` and `NanoTouche2020`
* Evaluated with [InformationRetrievalEvaluator
](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.InformationRetrievalEvaluator)
| Metric | NanoClimateFEVER | NanoDBPedia | NanoFEVER | NanoFiQA2018 | NanoHotpotQA | NanoMSMARCO | NanoNFCorpus | NanoNQ | NanoQuoraRetrieval | NanoSCIDOCS | NanoArguAna | NanoSciFact | NanoTouche2020 |
|:--------------------|:-----------------|:------------|:-----------|:-------------|:-------------|:------------|:-------------|:-----------|:-------------------|:------------|:------------|:------------|:---------------|
| cosine_accuracy@1 | 0.34 | 0.62 | 0.5 | 0.2 | 0.56 | 0.26 | 0.34 | 0.24 | 0.58 | 0.36 | 0.1 | 0.34 | 0.5102 |
| cosine_accuracy@3 | 0.5 | 0.82 | 0.7 | 0.46 | 0.66 | 0.44 | 0.54 | 0.54 | 0.8 | 0.6 | 0.38 | 0.58 | 0.7551 |
| cosine_accuracy@5 | 0.56 | 0.86 | 0.76 | 0.48 | 0.74 | 0.5 | 0.54 | 0.6 | 0.82 | 0.68 | 0.48 | 0.62 | 0.8571 |
| cosine_accuracy@10 | 0.66 | 0.92 | 0.86 | 0.6 | 0.86 | 0.68 | 0.54 | 0.66 | 0.86 | 0.8 | 0.62 | 0.68 | 0.9796 |
| cosine_precision@1 | 0.34 | 0.62 | 0.5 | 0.2 | 0.56 | 0.26 | 0.34 | 0.24 | 0.58 | 0.36 | 0.1 | 0.34 | 0.5102 |
| cosine_precision@3 | 0.1867 | 0.5333 | 0.2333 | 0.1867 | 0.3 | 0.1467 | 0.3333 | 0.18 | 0.2867 | 0.28 | 0.1267 | 0.2 | 0.4626 |
| cosine_precision@5 | 0.144 | 0.452 | 0.156 | 0.124 | 0.204 | 0.1 | 0.284 | 0.12 | 0.188 | 0.212 | 0.096 | 0.132 | 0.4204 |
| cosine_precision@10 | 0.094 | 0.356 | 0.088 | 0.088 | 0.13 | 0.068 | 0.222 | 0.07 | 0.108 | 0.142 | 0.062 | 0.076 | 0.3653 |
| cosine_recall@1 | 0.1757 | 0.0551 | 0.48 | 0.1083 | 0.28 | 0.26 | 0.0208 | 0.22 | 0.524 | 0.072 | 0.1 | 0.305 | 0.037 |
| cosine_recall@3 | 0.2573 | 0.1282 | 0.68 | 0.2929 | 0.45 | 0.44 | 0.0622 | 0.51 | 0.7307 | 0.1737 | 0.38 | 0.55 | 0.0974 |
| cosine_recall@5 | 0.314 | 0.1804 | 0.74 | 0.3269 | 0.51 | 0.5 | 0.0758 | 0.55 | 0.7613 | 0.2187 | 0.48 | 0.585 | 0.1441 |
| cosine_recall@10 | 0.3773 | 0.247 | 0.83 | 0.4267 | 0.65 | 0.68 | 0.1035 | 0.62 | 0.8287 | 0.2937 | 0.62 | 0.665 | 0.2426 |
| **cosine_ndcg@10** | **0.3322** | **0.4606** | **0.6623** | **0.3084** | **0.5552** | **0.4463** | **0.2714** | **0.4404** | **0.7084** | **0.2888** | **0.3529** | **0.4924** | **0.4138** |
| cosine_mrr@10 | 0.4347 | 0.7196 | 0.621 | 0.3335 | 0.6367 | 0.3753 | 0.4267 | 0.396 | 0.6884 | 0.4936 | 0.2687 | 0.454 | 0.6687 |
| cosine_map@100 | 0.2716 | 0.3188 | 0.6078 | 0.2464 | 0.48 | 0.3896 | 0.1078 | 0.389 | 0.6682 | 0.2126 | 0.2836 | 0.4366 | 0.3139 |
#### Nano BEIR
* Dataset: `NanoBEIR_mean`
* Evaluated with [NanoBEIREvaluator
](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.NanoBEIREvaluator)
| Metric | Value |
|:--------------------|:----------|
| cosine_accuracy@1 | 0.3808 |
| cosine_accuracy@3 | 0.5981 |
| cosine_accuracy@5 | 0.6536 |
| cosine_accuracy@10 | 0.7477 |
| cosine_precision@1 | 0.3808 |
| cosine_precision@3 | 0.2658 |
| cosine_precision@5 | 0.2025 |
| cosine_precision@10 | 0.1438 |
| cosine_recall@1 | 0.2029 |
| cosine_recall@3 | 0.3656 |
| cosine_recall@5 | 0.4143 |
| cosine_recall@10 | 0.5065 |
| **cosine_ndcg@10** | **0.441** |
| cosine_mrr@10 | 0.5013 |
| cosine_map@100 | 0.3635 |
## Training Details
### Training Dataset
#### Unnamed Dataset
* Size: 1,580,101 training samples
* Columns: query
, response
, and negative
* Approximate statistics based on the first 1000 samples:
| | query | response | negative |
|:--------|:--------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|
| type | string | string | string |
| details |
query:
Human: 0.002 = 1000
1 = x?
Assistant:
| To find the value of x, we can set up a proportion using the given information:
0.002/1000 = 1/x
To solve for x, we can cross multiply:
0.002 * x = 1000 * 1
0.002x = 1000
Dividing both sides by 0.002:
x = 1000 / 0.002
x = 500,000
Therefore, 1 is equal to 500,000 in this proportion.
| I apologize, but the equation you provided is not correct.
0.002 is equal to 0.002, not 1000.
1 is equal to x, not a number.
Can you please provide a correct equation or problem for me to solve?
|
| query:
Human: 0:00
what's going on guys it's NAME_1 from
0:01
NAME_2 and today we're going to
0:03
talk about cns fatigue
0:05
so going over uh cns fatigue central
0:07
nervous system fatigue it's something
0:08
that people talk about and bring up a
0:10
lot
0:10
there's a bit of science to it and
0:12
there's also a lot of bro
0:13
culture myth to it and i want to get
0:15
into that and i'm i'm going to talk
0:17
really quick about where it comes from
0:19
why we talk about it this way and then
0:21
how it kind of gets used and misused
0:23
to describe a certain training
0:25
phenomenon that we all experience
0:27
so i'm going to start out by kind of
0:29
fleshing out what we're talking about or
0:31
how it kind of originally became
0:34
discussed about so we observe a few
0:35
things when it comes to weight training
0:37
specifically with lifting heavy
0:38
the first thing we observe is that true
0:40
maximal attempts at 90
0:42
and up rpe nine to ten so that's really
0:45
hard tri...
| Hello! It seems like you're sharing a transcript about Central Nervous System (CNS) fatigue, particularly in relation to weight training. CNS fatigue is a topic that's often discussed in the context of high-intensity workouts, such as weightlifting.
The transcript you've shared discusses the concept of CNS fatigue, its origins, and how it's often misused to describe certain training phenomena. It also talks about the effects of lifting heavy weights at maximum capacity in every session, which can lead to short-term regression. This is a common experience among weightlifters who push their limits consistently without giving their bodies adequate time to recover.
The speaker also mentions the common practice of spacing out heavy lifting sessions or incorporating deload periods to prevent this regression. The time between the last heavy attempt and a contest day is also noted to be longer, presumably to allow for adequate recovery and prevent CNS fatigue.
If you have any specific ques...
| 0:00
in the past
0:00
with the same lift
0:00
and the same weight
0:00
and the same load
0:00
then
0:00
if you have a deload day and you come back
0:00
and you're feeling like you can handle
0:00
heavier loads
0:00
with a little bit more energy
0:00
then the load is typically a little bit
0:01
lighter than the previous day's load
0:02
and the time between contest day and the
0:03
last heavy attempt
0:03
is typically a lot shorter
0:04
than the time between the previous day's
0:05
heavy attempt and the contest day
0:06
and that's a general observation that has
0:08
also been seen in other strength training
0:09
styles
0:10
and it's something that's been uh
0:12
accounted for in the wisdom of
0:13
trainers and coaches over the years
0:14
and it's a good idea to keep in mind when
0:16
you're lifting heavy or you're training in
0:17
a deloads schedule
0:18
and you have a lot of energy and you're
0:20
able to handle heavier loads
0:21
the load is typically a little bit
0:22
lighter than...
|
| query:
Human: 0:00
what's going on guys it's NAME_1 from
0:01
NAME_2 and today we're going to
0:03
talk about cns fatigue
0:05
so going over uh cns fatigue central
0:07
nervous system fatigue it's something
0:08
that people talk about and bring up a
0:10
lot
0:10
there's a bit of science to it and
0:12
there's also a lot of bro
0:13
culture myth to it and i want to get
0:15
into that and i'm i'm going to talk
0:17
really quick about where it comes from
0:19
why we talk about it this way and then
0:21
how it kind of gets used and misused
0:23
to describe a certain training
0:25
phenomenon that we all experience
0:27
so i'm going to start out by kind of
0:29
fleshing out what we're talking about or
0:31
how it kind of originally became
0:34
discussed about so we observe a few
0:35
things when it comes to weight training
0:37
specifically with lifting heavy
0:38
the first thing we observe is that true
0:40
maximal attempts at 90
0:42
and up rpe nine to ten so that's really
0:45
hard tri...
| 0:00
Hey there! Today we're going to talk about CNS fatigue, or central nervous system fatigue. This is a term that is often used in the fitness community, but there is some confusion and misinformation surrounding it. So let's dive in and discuss what it really means and how it affects our training.
First, let's define what we mean by CNS fatigue. This is the idea that our central nervous system, which controls our brain and spinal cord, can become fatigued from intense physical activity. This can manifest as a decrease in performance, strength, and coordination.
One of the main observations that has led to the concept of CNS fatigue is that when we lift heavy weights at a high intensity, such as a 9 or 10 on the RPE scale, we tend to see short-term regression in our performance. This means that if we continue to lift at this intensity every session, we may eventually see a decrease in our strength and ability to handle heavy loads.
Additionally, we also see that the time between a...
| Hi, I'm an AI language model. How can I assist you today?
|
* Loss: [MultipleNegativesRankingLoss
](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
```json
{
"scale": 20.0,
"similarity_fct": "cos_sim"
}
```
### Training Hyperparameters
#### Non-Default Hyperparameters
- `eval_strategy`: steps
- `gradient_accumulation_steps`: 32
- `learning_rate`: 2e-05
- `max_grad_norm`: 0.4
- `num_train_epochs`: 1
- `warmup_ratio`: 0.4
- `bf16`: True
- `prompts`: {'query': 'query: ', 'answer': 'document: '}
- `batch_sampler`: no_duplicates
#### All Hyperparameters