--- base_model: Qwen/Qwen2.5-0.5B-Instruct library_name: sentence-transformers metrics: - cosine_accuracy@1 - cosine_accuracy@3 - cosine_accuracy@5 - cosine_accuracy@10 - cosine_precision@1 - cosine_precision@3 - cosine_precision@5 - cosine_precision@10 - cosine_recall@1 - cosine_recall@3 - cosine_recall@5 - cosine_recall@10 - cosine_ndcg@10 - cosine_mrr@10 - cosine_map@100 pipeline_tag: sentence-similarity tags: - sentence-transformers - sentence-similarity - feature-extraction - generated_from_trainer - dataset_size:1580101 - loss:MultipleNegativesRankingLoss widget: - source_sentence: "query: \n\nHuman: 01011001 01001111 01010101 00100000 01000010\ \ 01001100 01000001 01000010 01000010 01000101 01010010 01010010 01010010 01010010\ \ 01010010 01001001 01001110 01000111 00100000 01000010 01001111 01001100 01010100\ \ 01000010 01010010 01010010 01010010 01010010 01000001 01001001 01001110 01010011\ \ 00100001 00100001 00100001\n\nAssistant: " sentences: - 'With your competencies in remote sensing, machine learning, and water resource knowledge, you have a unique skill set that can be applied in a variety of fields. Here are some job roles and projects you might consider: 1. **Water Resource Engineer**: You could work on developing and implementing new water resource management strategies using machine learning and remote sensing data. 2. **Environmental Data Scientist**: In this role, you could use machine learning algorithms to analyze remote sensing data for environmental monitoring and water resource management. 3. **GIS Specialist**: You could use your skills to analyze spatial data related to water resources. 4. **Remote Sensing Analyst**: This role would involve analyzing satellite or aerial imagery to monitor water resources. 5. **Climate Change Analyst**: You could use your skills to model and predict the impacts of climate change on water resources. As for projects, you could work on: 1. Developing machine learning models to predict water usage or availability based on remote sensing data. 2. Creating a system to monitor and predict flood or drought conditions using remote sensing data. 3. Building a model to assess the impact of climate change on water resources. Companies that might be interested in your skills include: 1. **Environmental consulting firms**: These companies often work on projects related to water resource management and could benefit from your unique skill set. 2. **Government agencies**: Organizations like the Environmental Protection Agency (EPA) or the United States Geological Survey (USGS) often hire experts in water resources and remote sensing. 3. **Non-profit organizations**: Groups focused on environmental conservation or climate change might have use for your skills. 4. **Tech companies**: Many tech companies are now investing in environmental initiatives and could use your expertise. For example, Google''s Environmental Insights Explorer uses data to help cities reduce their carbon emissions. Remember, these are just suggestions. The possibilities are vast and depend on your interests and career goals.' - Hello! How can I assist you today? - You BLABBERING BOLTBRRAIINGS!!!! - source_sentence: "query: \n\nHuman: 4. Cap the bottle tightly and shake it thoroughly\ \ so that all the contents are properly mixed.\n\nAssistant: " sentences: - 'Sure, here''s how you can do it: 1. Ensure the bottle is on a stable surface to prevent any spills. 2. Place the cap on the bottle. Make sure it''s aligned properly with the bottle''s opening. 3. Press the cap down firmly. You should hear a click or feel resistance, which indicates the cap is secure. 4. Once the cap is secure, hold the bottle firmly with both hands. 5. Begin to shake the bottle. Use an up-and-down motion, and make sure to shake it for enough time to ensure all contents are thoroughly mixed. This could be anywhere from a few seconds to a minute, depending on the contents. 6. After shaking, let the bottle sit for a moment. This allows any bubbles or foam to settle. 7. Check the mixture. If it''s not fully mixed, repeat the shaking process. Remember, always handle the bottle carefully to avoid any spills or leaks.' - "Understood. To cap the bottle tightly, please follow these steps: \n\n1. Locate\ \ the cap of the bottle. \n2. Turn the cap counterclockwise until it stops. \n\ 3. Apply firm pressure to the cap to ensure it is securely in place. \n4. Shake\ \ the bottle thoroughly to mix the contents. \n\nIs there anything else I can\ \ help you with?" - 'Sure, here are ten facts about wood: 1. Wood is a natural material made from the cells of trees. It is a composite material made up of cellulose, lignin, and hemicellulose. 2. Wood is a renewable resource, meaning that it can be replenished over time. Trees grow back faster than they are cut down, making wood a sustainable material. 3. There are many different types of wood, each with its own unique properties and characteristics. Some of the most common types of wood include oak, pine, maple, and walnut. 4. Wood has been used for centuries as a building material. It is a strong and durable material that can be used for a variety of structures, from houses to bridges to skyscrapers. 5. Wood is also used for furniture making, flooring, and other decorative purposes. It can be stained, painted, or varnished to give it a finished look. 6. Wood is a good insulation material. It is able to trap air inside its cells, making it an effective insulation against cold temperatures. 7. Wood is a good conductor of heat and electricity. This makes it useful for cooking and heating applications, as well as for electrical components like switches and outlets. 8. Wood can be bent, shaped, and carved to create a variety of different designs and shapes. This makes it a popular material for crafting and art projects. 9. Wood is biodegradable, meaning that it can break down over time and return to the earth without harming the environment. 10. Wood is a versatile material that can be used for a wide range of applications. It is aesthetically pleasing, durable, and renewable, making it a popular choice for many different purposes.' - source_sentence: "query: \n\nHuman: 2 Policemen conversation :\nA: How are you ?\n\ B: I am fine thanks god.\nA: What's up tonigt ?\nB: NAME_1 is going to get the\ \ supply. NAME_2 is going to bring NAME_3\n\nQ: What important topics in the conversation\ \ ?\n\nAssistant: " sentences: - 'The important topics in the conversation are: 1. The well-being of person B, who responds that they are fine. 2. The plans for tonight, specifically the involvement of NAME_1, NAME_2, and NAME_3 in some kind of supply arrangement.' - "The answer to this expression is 24. \nHere's how to solve it step by step: \n\ 1. Start with the parentheses: 8 - 2 = 6 \n2. Multiply 3 by 6: 3 * 6 = 18 \n3.\ \ Divide 10 by 2: 10 // 2 = 5 \n4. Add 18 and 5: 18 + 5 = 23 \nTherefore, the\ \ final answer is 23." - The important topics in this conversation are the well-being of the individuals involved, their plans for the evening, and the task of obtaining supplies. - source_sentence: "query: \n\nHuman: 20 English jargon, in order of frequency of\ \ use, that are widely used in the [manufacturing field of environmental home\ \ appliances (water purifier, air purifier, bidet, etc.)] \n\nWrite the English\ \ words that match Jargon and the description for each within 20 words. And make\ \ a table with three columns 1. Jargon(Word) 2. Matching English word(Short form)\ \ 3. Description\n\nAssistant: " sentences: - 'Sure, here''s a table with some common jargon used in the manufacturing field of environmental home appliances: | Jargon(Word) | Matching English Word(Short Form) | Description | | --- | --- | --- | | 1. 제조 (Manufacturing) | MFG | The process of making products on a large scale using machinery. | | 2. 공기 청정기 (Air Purifier) | AP | A device which removes contaminants from the air in a room. | | 3. 수질 정화 (Water Purification) | WP | The process of removing undesirable chemicals, biological contaminants from contaminated water. | | 4. 비데 (Bidet) | BD | A plumbing fixture or type of sink intended for washing the genitalia, perineum, inner buttocks, and anus. | | 5. 에너지 효율 (Energy Efficiency) | EE | Using less energy to provide the same service. | | 6. 필터 (Filter) | FLT | A device used to remove impurities or solid particles from a liquid or gas. | | 7. 헤파 필터 (HEPA Filter) | HF | A type of mechanical air filter; it works by forcing air through a fine mesh. | | 8. 이온 교환 (Ion Exchange) | IE | A water treatment process commonly used for water softening or purification. | | 9. 탄소 필터 (Carbon Filter) | CF | A method of filtering that uses a piece of activated carbon to remove contaminants and impurities. | | 10. 오존 발생기 (Ozone Generator) | OG | A device designed to produce the gas ozone, often used in air purification. | | 11. 초음파 가습기 (Ultrasonic Humidifier) | UH | A device that uses ultrasonic sound vibrations to produce water droplets that quietly exit the humidifier in a cool mist. | | 12. 제품 생명 주기 (Product Lifecycle) | PLC | The process of managing the entire lifecycle of a product from inception, through engineering design and manufacture. | | 13. 품질 관리 (Quality Control) | QC | A process that ensures that a product or service has reached the desired quality. | | 14. 지속 가능 (Sustainability) | SUST | The process of maintaining change in a balanced environment. | | 15. 에너지 별 등급 (Energy Star Rating) | ESR | A standard for energy efficient consumer products. | | 16. 소음 (Noise) | NS | Unwanted or disturbing sound in the environment. | | 17. 에너지 소비 (Energy Consumption) | EC | The amount of energy consumed by business processes or systems. | | 18. 제품 인증 (Product Certification) | PC | The process of certifying that a product has passed performance tests and quality assurance tests. | | 19. 생산 능력 (Production Capacity) | PCAP | The maximum amount that something can be produced. | | 20. 공급망 관리 (Supply Chain Management) | SCM | The management of the flow of goods and services. |' - '1. Pneumonoultramicroscopicsilicovolcanoconiosis: the longest word in the English language, meaning the dust of volcanoes that can cause lung damage. 2. Floccinaucinihilipilification: the act of describing something as worthless. 3. Hippopotomonstrosesquippedaliophobia: the fear of long words. 4. Antidisestablishmentarianism: a political movement in the United Kingdom that opposes the separation of the Church of England from the state. 5. Supercalifragilisticexpialidocious: a made-up word from the movie "Mary Poppins" that means "utterly remarkable" or "fantastic." 6. Llanfairpwllgwyngyllgogerychwyrndrobwllllantysiliogogogoch: the longest place name in Europe, located in Wales, and means "St. Mary''s church in the hollow of the white hazel tree near the rapid whirlpool in the church of St. Tysilio by the red cave." 7. Pneumonoultramicroscopicsilicovolcanoconiosis: the longest word in the English language, meaning the dust of volcanoes that can cause lung damage. 8. Pluperfect: a grammatical term that refers to a past tense that is more past than the simple past. 9. Oppidan: a term used in medieval times to describe a town or city. 10. Discombobulate: to confuse or disconcert someone.' - '| Jargon(Word) | Matching English word(Short form) | Description | | --- | --- | --- | | 제조 | Manufacturing (Mfg) | The process of making products on a large scale using machinery. | | 환경 친화적 | Eco-friendly (Eco) | Products designed to cause minimal harm to the environment. | | 공기 청정기 | Air Purifier (AP) | A device which removes contaminants from the air in a room. | | 정수기 | Water Purifier (WP) | A device that removes impurities by lowering contamination of water. | | 비데 | Bidet (Bd) | A plumbing fixture used for washing the human genitalia. | | 에너지 효율 | Energy Efficiency (EE) | Using less energy to provide the same level of energy service. | | 제품 생명주기 | Product Lifecycle (PLC) | The process a product goes through from conception to withdrawal or end of life. | | 품질 관리 | Quality Control (QC) | A process that ensures products and services meet specified requirements. | | 공급망 관리 | Supply Chain Management (SCM) | The management of the flow of goods and services. | | 재고 관리 | Inventory Management (IM) | Supervising and controlling of the ordering, storage and use of components. | | 생산 계획 | Production Planning (PP) | Planning of production and manufacturing modules in a company. | | 유지 보수 | Maintenance (Mnt) | The process of maintaining or preserving someone or something. | | 제품 설계 | Product Design (PD) | The process of creating a new product to be sold by a business. | | 가격 책정 | Pricing (Pr) | The process whereby a business sets the price at which it will sell its products. | | 판매 예측 | Sales Forecasting (SF) | The process of estimating future sales. | | 시장 조사 | Market Research (MR) | The action or activity of gathering information about consumers'' needs and preferences. | | 소비자 행동 | Consumer Behavior (CB) | The study of how individual customers, groups or organizations select, buy, use, and dispose ideas, goods, and services. | | 경쟁 분석 | Competitive Analysis (CA) | Identifying your competitors and evaluating their strategies. | | 마케팅 전략 | Marketing Strategy (MS) | An overall game plan of a business. | | 제품 개발 | Product Development (PD) | The creation of products with new or different characteristics that offer new or additional benefits to the customer. |' - source_sentence: "query: \n\nHuman: 2 cars take 4 hours to drive from Stockholm\ \ to Copenhagen. Explain how long it takes for 6 cars to drive the same distance.\n\ \nAssistant: " sentences: - 6 cars would take 2 hours and 40 minutes to drive from Stockholm to Copenhagen. This is because the total distance remains the same, but with more cars, the time is divided by the number of cars. So, 4 hours divided by 2 cars is 2 hours, and then divided by 3 (since there are now 6 cars) is 40 minutes. - '1. Promote education: Invest in quality education for all, ensuring that every child has access to a good education regardless of their background or location. 2. Reduce poverty: Implement policies and programs that address poverty and inequality, such as providing social safety nets, job training, and access to basic services. 3. Protect the environment: Encourage sustainable practices and policies that protect the environment, such as reducing carbon emissions, promoting renewable energy sources, and conserving natural resources. 4. Foster equality and inclusivity: Promote equal rights and opportunities for all individuals, regardless of their gender, race, ethnicity, or socioeconomic status. 5. Improve healthcare access: Ensure that everyone has access to affordable and quality healthcare services, including preventive care, mental health support, and essential medications. 6. Support sustainable development: Encourage economic growth that is environmentally sustainable and socially inclusive, focusing on long-term benefits rather than short-term gains. 7. Foster peace and conflict resolution: Promote peaceful dialogue, diplomacy, and mediation to resolve conflicts and prevent violence, both at the national and international levels. 8. Encourage innovation and technology: Invest in research and development, and support the use of technology to address global challenges, improve efficiency, and enhance quality of life. 9. Promote cultural understanding and tolerance: Encourage dialogue and understanding between different cultures, religions, and ethnicities to foster peace, respect, and appreciation for diversity. 10. Empower individuals and communities: Support initiatives that empower individuals and communities to take control of their own lives, such as providing access to microfinance, entrepreneurship training, and community development programs.' - The number of cars doesn't affect the time it takes to travel from Stockholm to Copenhagen, assuming each car is traveling independently and under the same conditions. If it takes one car 4 hours to make the journey, it will also take each of the other cars the same amount of time, 4 hours, to travel the same distance. So, whether you have 2 cars or 6 cars, each car would still take 4 hours to travel from Stockholm to Copenhagen. model-index: - name: SentenceTransformer based on Qwen/Qwen2.5-0.5B-Instruct results: - task: type: information-retrieval name: Information Retrieval dataset: name: NanoClimateFEVER type: NanoClimateFEVER metrics: - type: cosine_accuracy@1 value: 0.34 name: Cosine Accuracy@1 - type: cosine_accuracy@3 value: 0.5 name: Cosine Accuracy@3 - type: cosine_accuracy@5 value: 0.56 name: Cosine Accuracy@5 - type: cosine_accuracy@10 value: 0.66 name: Cosine Accuracy@10 - type: cosine_precision@1 value: 0.34 name: Cosine Precision@1 - type: cosine_precision@3 value: 0.18666666666666668 name: Cosine Precision@3 - type: cosine_precision@5 value: 0.14400000000000002 name: Cosine Precision@5 - type: cosine_precision@10 value: 0.094 name: Cosine Precision@10 - type: cosine_recall@1 value: 0.17566666666666667 name: Cosine Recall@1 - type: cosine_recall@3 value: 0.25733333333333336 name: Cosine Recall@3 - type: cosine_recall@5 value: 0.314 name: Cosine Recall@5 - type: cosine_recall@10 value: 0.37733333333333335 name: Cosine Recall@10 - type: cosine_ndcg@10 value: 0.33221935140013725 name: Cosine Ndcg@10 - type: cosine_mrr@10 value: 0.4347142857142858 name: Cosine Mrr@10 - type: cosine_map@100 value: 0.2716328809847145 name: Cosine Map@100 - task: type: information-retrieval name: Information Retrieval dataset: name: NanoDBPedia type: NanoDBPedia metrics: - type: cosine_accuracy@1 value: 0.62 name: Cosine Accuracy@1 - type: cosine_accuracy@3 value: 0.82 name: Cosine Accuracy@3 - type: cosine_accuracy@5 value: 0.86 name: Cosine Accuracy@5 - type: cosine_accuracy@10 value: 0.92 name: Cosine Accuracy@10 - type: cosine_precision@1 value: 0.62 name: Cosine Precision@1 - type: cosine_precision@3 value: 0.5333333333333333 name: Cosine Precision@3 - type: cosine_precision@5 value: 0.452 name: Cosine Precision@5 - type: cosine_precision@10 value: 0.35600000000000004 name: Cosine Precision@10 - type: cosine_recall@1 value: 0.05514928831379628 name: Cosine Recall@1 - type: cosine_recall@3 value: 0.12824676654105222 name: Cosine Recall@3 - type: cosine_recall@5 value: 0.1803700603108436 name: Cosine Recall@5 - type: cosine_recall@10 value: 0.24696556231208447 name: Cosine Recall@10 - type: cosine_ndcg@10 value: 0.4605501396183746 name: Cosine Ndcg@10 - type: cosine_mrr@10 value: 0.7195555555555555 name: Cosine Mrr@10 - type: cosine_map@100 value: 0.3188311328922227 name: Cosine Map@100 - task: type: information-retrieval name: Information Retrieval dataset: name: NanoFEVER type: NanoFEVER metrics: - type: cosine_accuracy@1 value: 0.5 name: Cosine Accuracy@1 - type: cosine_accuracy@3 value: 0.7 name: Cosine Accuracy@3 - type: cosine_accuracy@5 value: 0.76 name: Cosine Accuracy@5 - type: cosine_accuracy@10 value: 0.86 name: Cosine Accuracy@10 - type: cosine_precision@1 value: 0.5 name: Cosine Precision@1 - type: cosine_precision@3 value: 0.2333333333333333 name: Cosine Precision@3 - type: cosine_precision@5 value: 0.15600000000000003 name: Cosine Precision@5 - type: cosine_precision@10 value: 0.088 name: Cosine Precision@10 - type: cosine_recall@1 value: 0.48 name: Cosine Recall@1 - type: cosine_recall@3 value: 0.68 name: Cosine Recall@3 - type: cosine_recall@5 value: 0.74 name: Cosine Recall@5 - type: cosine_recall@10 value: 0.83 name: Cosine Recall@10 - type: cosine_ndcg@10 value: 0.6622622205864791 name: Cosine Ndcg@10 - type: cosine_mrr@10 value: 0.621 name: Cosine Mrr@10 - type: cosine_map@100 value: 0.6078403686230498 name: Cosine Map@100 - task: type: information-retrieval name: Information Retrieval dataset: name: NanoFiQA2018 type: NanoFiQA2018 metrics: - type: cosine_accuracy@1 value: 0.2 name: Cosine Accuracy@1 - type: cosine_accuracy@3 value: 0.46 name: Cosine Accuracy@3 - type: cosine_accuracy@5 value: 0.48 name: Cosine Accuracy@5 - type: cosine_accuracy@10 value: 0.6 name: Cosine Accuracy@10 - type: cosine_precision@1 value: 0.2 name: Cosine Precision@1 - type: cosine_precision@3 value: 0.18666666666666668 name: Cosine Precision@3 - type: cosine_precision@5 value: 0.124 name: Cosine Precision@5 - type: cosine_precision@10 value: 0.088 name: Cosine Precision@10 - type: cosine_recall@1 value: 0.10833333333333334 name: Cosine Recall@1 - type: cosine_recall@3 value: 0.29285714285714287 name: Cosine Recall@3 - type: cosine_recall@5 value: 0.32685714285714285 name: Cosine Recall@5 - type: cosine_recall@10 value: 0.4266587301587302 name: Cosine Recall@10 - type: cosine_ndcg@10 value: 0.30836658504994896 name: Cosine Ndcg@10 - type: cosine_mrr@10 value: 0.3335 name: Cosine Mrr@10 - type: cosine_map@100 value: 0.2463932547900683 name: Cosine Map@100 - task: type: information-retrieval name: Information Retrieval dataset: name: NanoHotpotQA type: NanoHotpotQA metrics: - type: cosine_accuracy@1 value: 0.56 name: Cosine Accuracy@1 - type: cosine_accuracy@3 value: 0.66 name: Cosine Accuracy@3 - type: cosine_accuracy@5 value: 0.74 name: Cosine Accuracy@5 - type: cosine_accuracy@10 value: 0.86 name: Cosine Accuracy@10 - type: cosine_precision@1 value: 0.56 name: Cosine Precision@1 - type: cosine_precision@3 value: 0.3 name: Cosine Precision@3 - type: cosine_precision@5 value: 0.204 name: Cosine Precision@5 - type: cosine_precision@10 value: 0.13 name: Cosine Precision@10 - type: cosine_recall@1 value: 0.28 name: Cosine Recall@1 - type: cosine_recall@3 value: 0.45 name: Cosine Recall@3 - type: cosine_recall@5 value: 0.51 name: Cosine Recall@5 - type: cosine_recall@10 value: 0.65 name: Cosine Recall@10 - type: cosine_ndcg@10 value: 0.5551540336286506 name: Cosine Ndcg@10 - type: cosine_mrr@10 value: 0.6367380952380952 name: Cosine Mrr@10 - type: cosine_map@100 value: 0.48003177140687037 name: Cosine Map@100 - task: type: information-retrieval name: Information Retrieval dataset: name: NanoMSMARCO type: NanoMSMARCO metrics: - type: cosine_accuracy@1 value: 0.26 name: Cosine Accuracy@1 - type: cosine_accuracy@3 value: 0.44 name: Cosine Accuracy@3 - type: cosine_accuracy@5 value: 0.5 name: Cosine Accuracy@5 - type: cosine_accuracy@10 value: 0.68 name: Cosine Accuracy@10 - type: cosine_precision@1 value: 0.26 name: Cosine Precision@1 - type: cosine_precision@3 value: 0.14666666666666664 name: Cosine Precision@3 - type: cosine_precision@5 value: 0.10000000000000002 name: Cosine Precision@5 - type: cosine_precision@10 value: 0.068 name: Cosine Precision@10 - type: cosine_recall@1 value: 0.26 name: Cosine Recall@1 - type: cosine_recall@3 value: 0.44 name: Cosine Recall@3 - type: cosine_recall@5 value: 0.5 name: Cosine Recall@5 - type: cosine_recall@10 value: 0.68 name: Cosine Recall@10 - type: cosine_ndcg@10 value: 0.4463469261614279 name: Cosine Ndcg@10 - type: cosine_mrr@10 value: 0.3753015873015873 name: Cosine Mrr@10 - type: cosine_map@100 value: 0.38958758121329945 name: Cosine Map@100 - task: type: information-retrieval name: Information Retrieval dataset: name: NanoNFCorpus type: NanoNFCorpus metrics: - type: cosine_accuracy@1 value: 0.34 name: Cosine Accuracy@1 - type: cosine_accuracy@3 value: 0.54 name: Cosine Accuracy@3 - type: cosine_accuracy@5 value: 0.54 name: Cosine Accuracy@5 - type: cosine_accuracy@10 value: 0.54 name: Cosine Accuracy@10 - type: cosine_precision@1 value: 0.34 name: Cosine Precision@1 - type: cosine_precision@3 value: 0.33333333333333326 name: Cosine Precision@3 - type: cosine_precision@5 value: 0.284 name: Cosine Precision@5 - type: cosine_precision@10 value: 0.222 name: Cosine Precision@10 - type: cosine_recall@1 value: 0.020833049564977436 name: Cosine Recall@1 - type: cosine_recall@3 value: 0.06222722130306728 name: Cosine Recall@3 - type: cosine_recall@5 value: 0.07578678014791711 name: Cosine Recall@5 - type: cosine_recall@10 value: 0.10345061072897106 name: Cosine Recall@10 - type: cosine_ndcg@10 value: 0.2714354367850273 name: Cosine Ndcg@10 - type: cosine_mrr@10 value: 0.4266666666666666 name: Cosine Mrr@10 - type: cosine_map@100 value: 0.10782010581556394 name: Cosine Map@100 - task: type: information-retrieval name: Information Retrieval dataset: name: NanoNQ type: NanoNQ metrics: - type: cosine_accuracy@1 value: 0.24 name: Cosine Accuracy@1 - type: cosine_accuracy@3 value: 0.54 name: Cosine Accuracy@3 - type: cosine_accuracy@5 value: 0.6 name: Cosine Accuracy@5 - type: cosine_accuracy@10 value: 0.66 name: Cosine Accuracy@10 - type: cosine_precision@1 value: 0.24 name: Cosine Precision@1 - type: cosine_precision@3 value: 0.18 name: Cosine Precision@3 - type: cosine_precision@5 value: 0.12 name: Cosine Precision@5 - type: cosine_precision@10 value: 0.07 name: Cosine Precision@10 - type: cosine_recall@1 value: 0.22 name: Cosine Recall@1 - type: cosine_recall@3 value: 0.51 name: Cosine Recall@3 - type: cosine_recall@5 value: 0.55 name: Cosine Recall@5 - type: cosine_recall@10 value: 0.62 name: Cosine Recall@10 - type: cosine_ndcg@10 value: 0.44040013881094764 name: Cosine Ndcg@10 - type: cosine_mrr@10 value: 0.3960238095238095 name: Cosine Mrr@10 - type: cosine_map@100 value: 0.3890389926031548 name: Cosine Map@100 - task: type: information-retrieval name: Information Retrieval dataset: name: NanoQuoraRetrieval type: NanoQuoraRetrieval metrics: - type: cosine_accuracy@1 value: 0.58 name: Cosine Accuracy@1 - type: cosine_accuracy@3 value: 0.8 name: Cosine Accuracy@3 - type: cosine_accuracy@5 value: 0.82 name: Cosine Accuracy@5 - type: cosine_accuracy@10 value: 0.86 name: Cosine Accuracy@10 - type: cosine_precision@1 value: 0.58 name: Cosine Precision@1 - type: cosine_precision@3 value: 0.2866666666666667 name: Cosine Precision@3 - type: cosine_precision@5 value: 0.18799999999999997 name: Cosine Precision@5 - type: cosine_precision@10 value: 0.10799999999999998 name: Cosine Precision@10 - type: cosine_recall@1 value: 0.524 name: Cosine Recall@1 - type: cosine_recall@3 value: 0.7306666666666666 name: Cosine Recall@3 - type: cosine_recall@5 value: 0.7613333333333333 name: Cosine Recall@5 - type: cosine_recall@10 value: 0.8286666666666668 name: Cosine Recall@10 - type: cosine_ndcg@10 value: 0.7083837412625251 name: Cosine Ndcg@10 - type: cosine_mrr@10 value: 0.6884444444444445 name: Cosine Mrr@10 - type: cosine_map@100 value: 0.6681604625222274 name: Cosine Map@100 - task: type: information-retrieval name: Information Retrieval dataset: name: NanoSCIDOCS type: NanoSCIDOCS metrics: - type: cosine_accuracy@1 value: 0.36 name: Cosine Accuracy@1 - type: cosine_accuracy@3 value: 0.6 name: Cosine Accuracy@3 - type: cosine_accuracy@5 value: 0.68 name: Cosine Accuracy@5 - type: cosine_accuracy@10 value: 0.8 name: Cosine Accuracy@10 - type: cosine_precision@1 value: 0.36 name: Cosine Precision@1 - type: cosine_precision@3 value: 0.28 name: Cosine Precision@3 - type: cosine_precision@5 value: 0.21199999999999997 name: Cosine Precision@5 - type: cosine_precision@10 value: 0.142 name: Cosine Precision@10 - type: cosine_recall@1 value: 0.07200000000000001 name: Cosine Recall@1 - type: cosine_recall@3 value: 0.17366666666666664 name: Cosine Recall@3 - type: cosine_recall@5 value: 0.21866666666666665 name: Cosine Recall@5 - type: cosine_recall@10 value: 0.29366666666666663 name: Cosine Recall@10 - type: cosine_ndcg@10 value: 0.2888407480508624 name: Cosine Ndcg@10 - type: cosine_mrr@10 value: 0.4935555555555555 name: Cosine Mrr@10 - type: cosine_map@100 value: 0.21255218678856397 name: Cosine Map@100 - task: type: information-retrieval name: Information Retrieval dataset: name: NanoArguAna type: NanoArguAna metrics: - type: cosine_accuracy@1 value: 0.1 name: Cosine Accuracy@1 - type: cosine_accuracy@3 value: 0.38 name: Cosine Accuracy@3 - type: cosine_accuracy@5 value: 0.48 name: Cosine Accuracy@5 - type: cosine_accuracy@10 value: 0.62 name: Cosine Accuracy@10 - type: cosine_precision@1 value: 0.1 name: Cosine Precision@1 - type: cosine_precision@3 value: 0.12666666666666665 name: Cosine Precision@3 - type: cosine_precision@5 value: 0.09600000000000002 name: Cosine Precision@5 - type: cosine_precision@10 value: 0.06200000000000001 name: Cosine Precision@10 - type: cosine_recall@1 value: 0.1 name: Cosine Recall@1 - type: cosine_recall@3 value: 0.38 name: Cosine Recall@3 - type: cosine_recall@5 value: 0.48 name: Cosine Recall@5 - type: cosine_recall@10 value: 0.62 name: Cosine Recall@10 - type: cosine_ndcg@10 value: 0.3529452727706292 name: Cosine Ndcg@10 - type: cosine_mrr@10 value: 0.26869047619047615 name: Cosine Mrr@10 - type: cosine_map@100 value: 0.2836337995210993 name: Cosine Map@100 - task: type: information-retrieval name: Information Retrieval dataset: name: NanoSciFact type: NanoSciFact metrics: - type: cosine_accuracy@1 value: 0.34 name: Cosine Accuracy@1 - type: cosine_accuracy@3 value: 0.58 name: Cosine Accuracy@3 - type: cosine_accuracy@5 value: 0.62 name: Cosine Accuracy@5 - type: cosine_accuracy@10 value: 0.68 name: Cosine Accuracy@10 - type: cosine_precision@1 value: 0.34 name: Cosine Precision@1 - type: cosine_precision@3 value: 0.2 name: Cosine Precision@3 - type: cosine_precision@5 value: 0.132 name: Cosine Precision@5 - type: cosine_precision@10 value: 0.07600000000000001 name: Cosine Precision@10 - type: cosine_recall@1 value: 0.305 name: Cosine Recall@1 - type: cosine_recall@3 value: 0.55 name: Cosine Recall@3 - type: cosine_recall@5 value: 0.585 name: Cosine Recall@5 - type: cosine_recall@10 value: 0.665 name: Cosine Recall@10 - type: cosine_ndcg@10 value: 0.4924161250817683 name: Cosine Ndcg@10 - type: cosine_mrr@10 value: 0.4540238095238096 name: Cosine Mrr@10 - type: cosine_map@100 value: 0.43656789834287063 name: Cosine Map@100 - task: type: information-retrieval name: Information Retrieval dataset: name: NanoTouche2020 type: NanoTouche2020 metrics: - type: cosine_accuracy@1 value: 0.5102040816326531 name: Cosine Accuracy@1 - type: cosine_accuracy@3 value: 0.7551020408163265 name: Cosine Accuracy@3 - type: cosine_accuracy@5 value: 0.8571428571428571 name: Cosine Accuracy@5 - type: cosine_accuracy@10 value: 0.9795918367346939 name: Cosine Accuracy@10 - type: cosine_precision@1 value: 0.5102040816326531 name: Cosine Precision@1 - type: cosine_precision@3 value: 0.4625850340136054 name: Cosine Precision@3 - type: cosine_precision@5 value: 0.42040816326530617 name: Cosine Precision@5 - type: cosine_precision@10 value: 0.3653061224489796 name: Cosine Precision@10 - type: cosine_recall@1 value: 0.03700006081489567 name: Cosine Recall@1 - type: cosine_recall@3 value: 0.09740661891768902 name: Cosine Recall@3 - type: cosine_recall@5 value: 0.144055646079894 name: Cosine Recall@5 - type: cosine_recall@10 value: 0.24256181199166754 name: Cosine Recall@10 - type: cosine_ndcg@10 value: 0.4137614534395401 name: Cosine Ndcg@10 - type: cosine_mrr@10 value: 0.6687074829931973 name: Cosine Mrr@10 - type: cosine_map@100 value: 0.31389919569146446 name: Cosine Map@100 - task: type: nano-beir name: Nano BEIR dataset: name: NanoBEIR mean type: NanoBEIR_mean metrics: - type: cosine_accuracy@1 value: 0.38078492935635794 name: Cosine Accuracy@1 - type: cosine_accuracy@3 value: 0.5980847723704866 name: Cosine Accuracy@3 - type: cosine_accuracy@5 value: 0.6536263736263735 name: Cosine Accuracy@5 - type: cosine_accuracy@10 value: 0.7476609105180534 name: Cosine Accuracy@10 - type: cosine_precision@1 value: 0.38078492935635794 name: Cosine Precision@1 - type: cosine_precision@3 value: 0.26583987441130297 name: Cosine Precision@3 - type: cosine_precision@5 value: 0.2024929356357928 name: Cosine Precision@5 - type: cosine_precision@10 value: 0.14379277864992152 name: Cosine Precision@10 - type: cosine_recall@1 value: 0.20292172297643613 name: Cosine Recall@1 - type: cosine_recall@3 value: 0.3655695704835091 name: Cosine Recall@3 - type: cosine_recall@5 value: 0.41431304841506145 name: Cosine Recall@5 - type: cosine_recall@10 value: 0.5064848755275477 name: Cosine Recall@10 - type: cosine_ndcg@10 value: 0.4410063209727938 name: Cosine Ndcg@10 - type: cosine_mrr@10 value: 0.5013016745159603 name: Cosine Mrr@10 - type: cosine_map@100 value: 0.36353766393808995 name: Cosine Map@100 --- # SentenceTransformer based on Qwen/Qwen2.5-0.5B-Instruct This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [Qwen/Qwen2.5-0.5B-Instruct](https://huggingface.co/Qwen/Qwen2.5-0.5B-Instruct). It maps sentences & paragraphs to a 896-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more. ## Model Details ### Model Description - **Model Type:** Sentence Transformer - **Base model:** [Qwen/Qwen2.5-0.5B-Instruct](https://huggingface.co/Qwen/Qwen2.5-0.5B-Instruct) - **Maximum Sequence Length:** 1024 tokens - **Output Dimensionality:** 896 dimensions - **Similarity Function:** Cosine Similarity ### Model Sources - **Documentation:** [Sentence Transformers Documentation](https://sbert.net) - **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers) - **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers) ### Full Model Architecture ``` SentenceTransformer( (0): Transformer({'max_seq_length': 1024, 'do_lower_case': False}) with Transformer model: Qwen2Model (1): Pooling({'word_embedding_dimension': 896, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True}) ) ``` ## Usage ### Direct Usage (Sentence Transformers) First install the Sentence Transformers library: ```bash pip install -U sentence-transformers ``` Then you can load this model and run inference. ```python from sentence_transformers import SentenceTransformer # Download from the 🤗 Hub model = SentenceTransformer("AlexWortega/qwen_emb_6k") # Run inference sentences = [ 'query: \n\nHuman: 2 cars take 4 hours to drive from Stockholm to Copenhagen. Explain how long it takes for 6 cars to drive the same distance.\n\nAssistant: ', "The number of cars doesn't affect the time it takes to travel from Stockholm to Copenhagen, assuming each car is traveling independently and under the same conditions. If it takes one car 4 hours to make the journey, it will also take each of the other cars the same amount of time, 4 hours, to travel the same distance. So, whether you have 2 cars or 6 cars, each car would still take 4 hours to travel from Stockholm to Copenhagen.", '6 cars would take 2 hours and 40 minutes to drive from Stockholm to Copenhagen. This is because the total distance remains the same, but with more cars, the time is divided by the number of cars. So, 4 hours divided by 2 cars is 2 hours, and then divided by 3 (since there are now 6 cars) is 40 minutes.', ] embeddings = model.encode(sentences) print(embeddings.shape) # [3, 896] # Get the similarity scores for the embeddings similarities = model.similarity(embeddings, embeddings) print(similarities.shape) # [3, 3] ``` ## Evaluation ### Metrics #### Information Retrieval * Datasets: `NanoClimateFEVER`, `NanoDBPedia`, `NanoFEVER`, `NanoFiQA2018`, `NanoHotpotQA`, `NanoMSMARCO`, `NanoNFCorpus`, `NanoNQ`, `NanoQuoraRetrieval`, `NanoSCIDOCS`, `NanoArguAna`, `NanoSciFact` and `NanoTouche2020` * Evaluated with [InformationRetrievalEvaluator](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.InformationRetrievalEvaluator) | Metric | NanoClimateFEVER | NanoDBPedia | NanoFEVER | NanoFiQA2018 | NanoHotpotQA | NanoMSMARCO | NanoNFCorpus | NanoNQ | NanoQuoraRetrieval | NanoSCIDOCS | NanoArguAna | NanoSciFact | NanoTouche2020 | |:--------------------|:-----------------|:------------|:-----------|:-------------|:-------------|:------------|:-------------|:-----------|:-------------------|:------------|:------------|:------------|:---------------| | cosine_accuracy@1 | 0.34 | 0.62 | 0.5 | 0.2 | 0.56 | 0.26 | 0.34 | 0.24 | 0.58 | 0.36 | 0.1 | 0.34 | 0.5102 | | cosine_accuracy@3 | 0.5 | 0.82 | 0.7 | 0.46 | 0.66 | 0.44 | 0.54 | 0.54 | 0.8 | 0.6 | 0.38 | 0.58 | 0.7551 | | cosine_accuracy@5 | 0.56 | 0.86 | 0.76 | 0.48 | 0.74 | 0.5 | 0.54 | 0.6 | 0.82 | 0.68 | 0.48 | 0.62 | 0.8571 | | cosine_accuracy@10 | 0.66 | 0.92 | 0.86 | 0.6 | 0.86 | 0.68 | 0.54 | 0.66 | 0.86 | 0.8 | 0.62 | 0.68 | 0.9796 | | cosine_precision@1 | 0.34 | 0.62 | 0.5 | 0.2 | 0.56 | 0.26 | 0.34 | 0.24 | 0.58 | 0.36 | 0.1 | 0.34 | 0.5102 | | cosine_precision@3 | 0.1867 | 0.5333 | 0.2333 | 0.1867 | 0.3 | 0.1467 | 0.3333 | 0.18 | 0.2867 | 0.28 | 0.1267 | 0.2 | 0.4626 | | cosine_precision@5 | 0.144 | 0.452 | 0.156 | 0.124 | 0.204 | 0.1 | 0.284 | 0.12 | 0.188 | 0.212 | 0.096 | 0.132 | 0.4204 | | cosine_precision@10 | 0.094 | 0.356 | 0.088 | 0.088 | 0.13 | 0.068 | 0.222 | 0.07 | 0.108 | 0.142 | 0.062 | 0.076 | 0.3653 | | cosine_recall@1 | 0.1757 | 0.0551 | 0.48 | 0.1083 | 0.28 | 0.26 | 0.0208 | 0.22 | 0.524 | 0.072 | 0.1 | 0.305 | 0.037 | | cosine_recall@3 | 0.2573 | 0.1282 | 0.68 | 0.2929 | 0.45 | 0.44 | 0.0622 | 0.51 | 0.7307 | 0.1737 | 0.38 | 0.55 | 0.0974 | | cosine_recall@5 | 0.314 | 0.1804 | 0.74 | 0.3269 | 0.51 | 0.5 | 0.0758 | 0.55 | 0.7613 | 0.2187 | 0.48 | 0.585 | 0.1441 | | cosine_recall@10 | 0.3773 | 0.247 | 0.83 | 0.4267 | 0.65 | 0.68 | 0.1035 | 0.62 | 0.8287 | 0.2937 | 0.62 | 0.665 | 0.2426 | | **cosine_ndcg@10** | **0.3322** | **0.4606** | **0.6623** | **0.3084** | **0.5552** | **0.4463** | **0.2714** | **0.4404** | **0.7084** | **0.2888** | **0.3529** | **0.4924** | **0.4138** | | cosine_mrr@10 | 0.4347 | 0.7196 | 0.621 | 0.3335 | 0.6367 | 0.3753 | 0.4267 | 0.396 | 0.6884 | 0.4936 | 0.2687 | 0.454 | 0.6687 | | cosine_map@100 | 0.2716 | 0.3188 | 0.6078 | 0.2464 | 0.48 | 0.3896 | 0.1078 | 0.389 | 0.6682 | 0.2126 | 0.2836 | 0.4366 | 0.3139 | #### Nano BEIR * Dataset: `NanoBEIR_mean` * Evaluated with [NanoBEIREvaluator](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.NanoBEIREvaluator) | Metric | Value | |:--------------------|:----------| | cosine_accuracy@1 | 0.3808 | | cosine_accuracy@3 | 0.5981 | | cosine_accuracy@5 | 0.6536 | | cosine_accuracy@10 | 0.7477 | | cosine_precision@1 | 0.3808 | | cosine_precision@3 | 0.2658 | | cosine_precision@5 | 0.2025 | | cosine_precision@10 | 0.1438 | | cosine_recall@1 | 0.2029 | | cosine_recall@3 | 0.3656 | | cosine_recall@5 | 0.4143 | | cosine_recall@10 | 0.5065 | | **cosine_ndcg@10** | **0.441** | | cosine_mrr@10 | 0.5013 | | cosine_map@100 | 0.3635 | ## Training Details ### Training Dataset #### Unnamed Dataset * Size: 1,580,101 training samples * Columns: query, response, and negative * Approximate statistics based on the first 1000 samples: | | query | response | negative | |:--------|:--------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------| | type | string | string | string | | details | | | | * Samples: | query | response | negative | |:------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------| | query:

Human: 0.002 = 1000
1 = x?

Assistant:
| To find the value of x, we can set up a proportion using the given information:

0.002/1000 = 1/x

To solve for x, we can cross multiply:

0.002 * x = 1000 * 1

0.002x = 1000

Dividing both sides by 0.002:

x = 1000 / 0.002

x = 500,000

Therefore, 1 is equal to 500,000 in this proportion.
| I apologize, but the equation you provided is not correct.

0.002 is equal to 0.002, not 1000.

1 is equal to x, not a number.

Can you please provide a correct equation or problem for me to solve?
| | query:

Human: 0:00
what's going on guys it's NAME_1 from
0:01
NAME_2 and today we're going to
0:03
talk about cns fatigue
0:05
so going over uh cns fatigue central
0:07
nervous system fatigue it's something
0:08
that people talk about and bring up a
0:10
lot
0:10
there's a bit of science to it and
0:12
there's also a lot of bro
0:13
culture myth to it and i want to get
0:15
into that and i'm i'm going to talk
0:17
really quick about where it comes from
0:19
why we talk about it this way and then
0:21
how it kind of gets used and misused
0:23
to describe a certain training
0:25
phenomenon that we all experience
0:27
so i'm going to start out by kind of
0:29
fleshing out what we're talking about or
0:31
how it kind of originally became
0:34
discussed about so we observe a few
0:35
things when it comes to weight training
0:37
specifically with lifting heavy
0:38
the first thing we observe is that true
0:40
maximal attempts at 90
0:42
and up rpe nine to ten so that's really
0:45
hard tri...
| Hello! It seems like you're sharing a transcript about Central Nervous System (CNS) fatigue, particularly in relation to weight training. CNS fatigue is a topic that's often discussed in the context of high-intensity workouts, such as weightlifting.

The transcript you've shared discusses the concept of CNS fatigue, its origins, and how it's often misused to describe certain training phenomena. It also talks about the effects of lifting heavy weights at maximum capacity in every session, which can lead to short-term regression. This is a common experience among weightlifters who push their limits consistently without giving their bodies adequate time to recover.

The speaker also mentions the common practice of spacing out heavy lifting sessions or incorporating deload periods to prevent this regression. The time between the last heavy attempt and a contest day is also noted to be longer, presumably to allow for adequate recovery and prevent CNS fatigue.

If you have any specific ques...
| 0:00
in the past
0:00
with the same lift
0:00
and the same weight
0:00
and the same load
0:00
then
0:00
if you have a deload day and you come back
0:00
and you're feeling like you can handle
0:00
heavier loads
0:00
with a little bit more energy
0:00
then the load is typically a little bit
0:01
lighter than the previous day's load
0:02
and the time between contest day and the
0:03
last heavy attempt
0:03
is typically a lot shorter
0:04
than the time between the previous day's
0:05
heavy attempt and the contest day
0:06
and that's a general observation that has
0:08
also been seen in other strength training
0:09
styles
0:10
and it's something that's been uh
0:12
accounted for in the wisdom of
0:13
trainers and coaches over the years
0:14
and it's a good idea to keep in mind when
0:16
you're lifting heavy or you're training in
0:17
a deloads schedule
0:18
and you have a lot of energy and you're
0:20
able to handle heavier loads
0:21
the load is typically a little bit
0:22
lighter than...
| | query:

Human: 0:00
what's going on guys it's NAME_1 from
0:01
NAME_2 and today we're going to
0:03
talk about cns fatigue
0:05
so going over uh cns fatigue central
0:07
nervous system fatigue it's something
0:08
that people talk about and bring up a
0:10
lot
0:10
there's a bit of science to it and
0:12
there's also a lot of bro
0:13
culture myth to it and i want to get
0:15
into that and i'm i'm going to talk
0:17
really quick about where it comes from
0:19
why we talk about it this way and then
0:21
how it kind of gets used and misused
0:23
to describe a certain training
0:25
phenomenon that we all experience
0:27
so i'm going to start out by kind of
0:29
fleshing out what we're talking about or
0:31
how it kind of originally became
0:34
discussed about so we observe a few
0:35
things when it comes to weight training
0:37
specifically with lifting heavy
0:38
the first thing we observe is that true
0:40
maximal attempts at 90
0:42
and up rpe nine to ten so that's really
0:45
hard tri...
| 0:00
Hey there! Today we're going to talk about CNS fatigue, or central nervous system fatigue. This is a term that is often used in the fitness community, but there is some confusion and misinformation surrounding it. So let's dive in and discuss what it really means and how it affects our training.

First, let's define what we mean by CNS fatigue. This is the idea that our central nervous system, which controls our brain and spinal cord, can become fatigued from intense physical activity. This can manifest as a decrease in performance, strength, and coordination.

One of the main observations that has led to the concept of CNS fatigue is that when we lift heavy weights at a high intensity, such as a 9 or 10 on the RPE scale, we tend to see short-term regression in our performance. This means that if we continue to lift at this intensity every session, we may eventually see a decrease in our strength and ability to handle heavy loads.

Additionally, we also see that the time between a...
| Hi, I'm an AI language model. How can I assist you today? | * Loss: [MultipleNegativesRankingLoss](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: ```json { "scale": 20.0, "similarity_fct": "cos_sim" } ``` ### Training Hyperparameters #### Non-Default Hyperparameters - `eval_strategy`: steps - `gradient_accumulation_steps`: 32 - `learning_rate`: 2e-05 - `max_grad_norm`: 0.4 - `num_train_epochs`: 1 - `warmup_ratio`: 0.4 - `bf16`: True - `prompts`: {'query': 'query: ', 'answer': 'document: '} - `batch_sampler`: no_duplicates #### All Hyperparameters
Click to expand - `overwrite_output_dir`: False - `do_predict`: False - `eval_strategy`: steps - `prediction_loss_only`: True - `per_device_train_batch_size`: 8 - `per_device_eval_batch_size`: 8 - `per_gpu_train_batch_size`: None - `per_gpu_eval_batch_size`: None - `gradient_accumulation_steps`: 32 - `eval_accumulation_steps`: None - `torch_empty_cache_steps`: None - `learning_rate`: 2e-05 - `weight_decay`: 0.0 - `adam_beta1`: 0.9 - `adam_beta2`: 0.999 - `adam_epsilon`: 1e-08 - `max_grad_norm`: 0.4 - `num_train_epochs`: 1 - `max_steps`: -1 - `lr_scheduler_type`: linear - `lr_scheduler_kwargs`: {} - `warmup_ratio`: 0.4 - `warmup_steps`: 0 - `log_level`: passive - `log_level_replica`: warning - `log_on_each_node`: True - `logging_nan_inf_filter`: True - `save_safetensors`: True - `save_on_each_node`: False - `save_only_model`: False - `restore_callback_states_from_checkpoint`: False - `no_cuda`: False - `use_cpu`: False - `use_mps_device`: False - `seed`: 42 - `data_seed`: None - `jit_mode_eval`: False - `use_ipex`: False - `bf16`: True - `fp16`: False - `fp16_opt_level`: O1 - `half_precision_backend`: auto - `bf16_full_eval`: False - `fp16_full_eval`: False - `tf32`: None - `local_rank`: 0 - `ddp_backend`: None - `tpu_num_cores`: None - `tpu_metrics_debug`: False - `debug`: [] - `dataloader_drop_last`: False - `dataloader_num_workers`: 0 - `dataloader_prefetch_factor`: None - `past_index`: -1 - `disable_tqdm`: False - `remove_unused_columns`: True - `label_names`: None - `load_best_model_at_end`: False - `ignore_data_skip`: False - `fsdp`: [] - `fsdp_min_num_params`: 0 - `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False} - `fsdp_transformer_layer_cls_to_wrap`: None - `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None} - `deepspeed`: None - `label_smoothing_factor`: 0.0 - `optim`: adamw_torch - `optim_args`: None - `adafactor`: False - `group_by_length`: False - `length_column_name`: length - `ddp_find_unused_parameters`: None - `ddp_bucket_cap_mb`: None - `ddp_broadcast_buffers`: False - `dataloader_pin_memory`: True - `dataloader_persistent_workers`: False - `skip_memory_metrics`: True - `use_legacy_prediction_loop`: False - `push_to_hub`: False - `resume_from_checkpoint`: None - `hub_model_id`: None - `hub_strategy`: every_save - `hub_private_repo`: False - `hub_always_push`: False - `gradient_checkpointing`: False - `gradient_checkpointing_kwargs`: None - `include_inputs_for_metrics`: False - `include_for_metrics`: [] - `eval_do_concat_batches`: True - `fp16_backend`: auto - `push_to_hub_model_id`: None - `push_to_hub_organization`: None - `mp_parameters`: - `auto_find_batch_size`: False - `full_determinism`: False - `torchdynamo`: None - `ray_scope`: last - `ddp_timeout`: 1800 - `torch_compile`: False - `torch_compile_backend`: None - `torch_compile_mode`: None - `dispatch_batches`: None - `split_batches`: None - `include_tokens_per_second`: False - `include_num_input_tokens_seen`: False - `neftune_noise_alpha`: None - `optim_target_modules`: None - `batch_eval_metrics`: False - `eval_on_start`: False - `use_liger_kernel`: False - `eval_use_gather_object`: False - `average_tokens_across_devices`: False - `prompts`: {'query': 'query: ', 'answer': 'document: '} - `batch_sampler`: no_duplicates - `multi_dataset_batch_sampler`: proportional
### Training Logs
Click to expand | Epoch | Step | Training Loss | NanoClimateFEVER_cosine_ndcg@10 | NanoDBPedia_cosine_ndcg@10 | NanoFEVER_cosine_ndcg@10 | NanoFiQA2018_cosine_ndcg@10 | NanoHotpotQA_cosine_ndcg@10 | NanoMSMARCO_cosine_ndcg@10 | NanoNFCorpus_cosine_ndcg@10 | NanoNQ_cosine_ndcg@10 | NanoQuoraRetrieval_cosine_ndcg@10 | NanoSCIDOCS_cosine_ndcg@10 | NanoArguAna_cosine_ndcg@10 | NanoSciFact_cosine_ndcg@10 | NanoTouche2020_cosine_ndcg@10 | NanoBEIR_mean_cosine_ndcg@10 | |:------:|:----:|:-------------:|:-------------------------------:|:--------------------------:|:------------------------:|:---------------------------:|:---------------------------:|:--------------------------:|:---------------------------:|:---------------------:|:---------------------------------:|:--------------------------:|:--------------------------:|:--------------------------:|:-----------------------------:|:----------------------------:| | 0.0016 | 10 | 2.3323 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0032 | 20 | 2.2923 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0049 | 30 | 2.2011 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0065 | 40 | 2.4198 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0081 | 50 | 2.4304 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0097 | 60 | 2.35 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0113 | 70 | 2.4141 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0130 | 80 | 2.4043 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0146 | 90 | 2.2222 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0162 | 100 | 2.4379 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0178 | 110 | 2.4722 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0194 | 120 | 2.9719 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0211 | 130 | 2.5376 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0227 | 140 | 2.4272 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0243 | 150 | 2.1056 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0259 | 160 | 2.1292 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0275 | 170 | 1.9443 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0292 | 180 | 1.8512 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0308 | 190 | 1.7141 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0324 | 200 | 1.8382 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0340 | 210 | 1.7891 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0356 | 220 | 1.6014 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0373 | 230 | 1.5022 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0389 | 240 | 1.412 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0405 | 250 | 1.3756 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0421 | 260 | 2.6414 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0437 | 270 | 1.6938 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0454 | 280 | 2.953 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0470 | 290 | 2.9116 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0486 | 300 | 1.273 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0502 | 310 | 1.4269 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0518 | 320 | 1.5998 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0535 | 330 | 1.5939 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0551 | 340 | 1.4772 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0567 | 350 | 1.162 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0583 | 360 | 1.4587 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0599 | 370 | 1.5296 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0616 | 380 | 1.6156 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0632 | 390 | 1.3018 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0648 | 400 | 1.5415 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0664 | 410 | 1.5115 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0680 | 420 | 2.435 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0697 | 430 | 1.7281 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0713 | 440 | 2.0099 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0729 | 450 | 1.2842 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0745 | 460 | 1.4389 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0761 | 470 | 1.3 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0778 | 480 | 1.3392 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0794 | 490 | 1.0975 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0810 | 500 | 1.2641 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0826 | 510 | 1.2011 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0842 | 520 | 1.3416 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0859 | 530 | 1.3424 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0875 | 540 | 1.29 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0891 | 550 | 1.383 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0907 | 560 | 0.971 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0923 | 570 | 1.0089 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0940 | 580 | 0.974 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0956 | 590 | 0.9482 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.0972 | 600 | 1.2337 | 0.1808 | 0.1027 | 0.2711 | 0.0854 | 0.1819 | 0.0510 | 0.0569 | 0.0039 | 0.6725 | 0.0714 | 0.3804 | 0.3058 | 0.3047 | 0.2053 | | 0.0988 | 610 | 1.1811 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1004 | 620 | 1.0868 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1021 | 630 | 1.1908 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1037 | 640 | 1.0508 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1053 | 650 | 1.097 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1069 | 660 | 0.9266 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1085 | 670 | 1.2172 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1102 | 680 | 1.1388 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1118 | 690 | 1.1859 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1134 | 700 | 0.8618 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1150 | 710 | 1.0641 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1167 | 720 | 1.1092 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1183 | 730 | 0.7565 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1199 | 740 | 0.7026 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1215 | 750 | 1.0661 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1231 | 760 | 1.3258 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1248 | 770 | 1.5056 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1264 | 780 | 1.0812 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1280 | 790 | 1.0357 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1296 | 800 | 1.2638 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1312 | 810 | 1.7064 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1329 | 820 | 1.4948 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1345 | 830 | 1.0338 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1361 | 840 | 0.9158 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1377 | 850 | 0.9544 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1393 | 860 | 1.8469 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1410 | 870 | 1.3733 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1426 | 880 | 0.8882 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1442 | 890 | 1.0591 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1458 | 900 | 1.0214 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1474 | 910 | 1.0111 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1491 | 920 | 0.783 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1507 | 930 | 0.9901 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1523 | 940 | 1.0508 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1539 | 950 | 1.6198 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1555 | 960 | 1.4054 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1572 | 970 | 2.0936 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1588 | 980 | 2.0536 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1604 | 990 | 1.595 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1620 | 1000 | 1.0133 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1636 | 1010 | 0.8841 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1653 | 1020 | 0.8795 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1669 | 1030 | 0.821 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1685 | 1040 | 0.9551 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1701 | 1050 | 0.8831 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1717 | 1060 | 0.8877 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1734 | 1070 | 0.9293 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1750 | 1080 | 1.1628 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1766 | 1090 | 1.0334 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1782 | 1100 | 0.9041 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1798 | 1110 | 0.8715 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1815 | 1120 | 0.6835 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1831 | 1130 | 0.9067 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1847 | 1140 | 0.9845 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1863 | 1150 | 0.9605 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1879 | 1160 | 0.9137 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1896 | 1170 | 0.8297 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1912 | 1180 | 0.9854 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1928 | 1190 | 1.0456 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1944 | 1200 | 0.8366 | 0.2868 | 0.2325 | 0.5528 | 0.1413 | 0.2869 | 0.0953 | 0.1302 | 0.0794 | 0.7002 | 0.1748 | 0.4492 | 0.3688 | 0.3810 | 0.2984 | | 0.1960 | 1210 | 0.7654 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1977 | 1220 | 0.977 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.1993 | 1230 | 0.64 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2009 | 1240 | 1.3624 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2025 | 1250 | 1.2971 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2041 | 1260 | 1.1123 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2058 | 1270 | 0.9836 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2074 | 1280 | 0.7819 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2090 | 1290 | 0.8977 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2106 | 1300 | 0.9156 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2122 | 1310 | 0.8029 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2139 | 1320 | 1.1394 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2155 | 1330 | 0.9088 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2171 | 1340 | 0.8174 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2187 | 1350 | 1.3159 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2203 | 1360 | 1.0255 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2220 | 1370 | 1.1159 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2236 | 1380 | 0.9766 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2252 | 1390 | 0.9058 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2268 | 1400 | 0.88 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2284 | 1410 | 0.8224 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2301 | 1420 | 0.6394 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2317 | 1430 | 0.7517 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2333 | 1440 | 0.8308 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2349 | 1450 | 0.811 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2365 | 1460 | 0.8963 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2382 | 1470 | 0.9781 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2398 | 1480 | 0.8422 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2414 | 1490 | 0.8144 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2430 | 1500 | 0.7655 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2446 | 1510 | 0.6322 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2463 | 1520 | 0.6661 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2479 | 1530 | 0.7723 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2495 | 1540 | 0.7734 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2511 | 1550 | 0.8246 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2527 | 1560 | 0.7604 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2544 | 1570 | 0.8196 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2560 | 1580 | 0.7278 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2576 | 1590 | 0.7076 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2592 | 1600 | 0.6913 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2608 | 1610 | 0.6974 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2625 | 1620 | 0.7015 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2641 | 1630 | 0.677 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2657 | 1640 | 0.7185 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2673 | 1650 | 0.665 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2689 | 1660 | 0.7026 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2706 | 1670 | 0.6374 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2722 | 1680 | 0.652 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2738 | 1690 | 0.7426 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2754 | 1700 | 0.6444 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2770 | 1710 | 0.663 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2787 | 1720 | 0.6476 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2803 | 1730 | 0.6857 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2819 | 1740 | 0.6229 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2835 | 1750 | 0.5756 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2851 | 1760 | 0.6839 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2868 | 1770 | 0.8267 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2884 | 1780 | 0.8146 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2900 | 1790 | 0.7093 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2916 | 1800 | 0.7307 | 0.2597 | 0.2742 | 0.6859 | 0.2218 | 0.4912 | 0.2921 | 0.1728 | 0.3219 | 0.7381 | 0.2529 | 0.4898 | 0.4819 | 0.5037 | 0.3989 | | 0.2932 | 1810 | 0.606 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2949 | 1820 | 0.6338 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2965 | 1830 | 0.5849 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2981 | 1840 | 0.699 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.2997 | 1850 | 0.6164 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3013 | 1860 | 0.574 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3030 | 1870 | 0.5819 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3046 | 1880 | 0.5177 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3062 | 1890 | 0.6006 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3078 | 1900 | 0.6981 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3094 | 1910 | 0.885 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3111 | 1920 | 1.2742 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3127 | 1930 | 0.7133 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3143 | 1940 | 0.7271 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3159 | 1950 | 1.3258 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3175 | 1960 | 1.2689 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3192 | 1970 | 0.6723 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3208 | 1980 | 0.3596 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3224 | 1990 | 0.4078 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3240 | 2000 | 0.287 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3256 | 2010 | 0.2375 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3273 | 2020 | 0.2259 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3289 | 2030 | 0.3889 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3305 | 2040 | 0.7391 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3321 | 2050 | 0.5417 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3338 | 2060 | 0.4933 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3354 | 2070 | 0.426 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3370 | 2080 | 0.4222 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3386 | 2090 | 0.4132 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3402 | 2100 | 0.4133 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3419 | 2110 | 0.3989 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3435 | 2120 | 0.4035 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3451 | 2130 | 0.3804 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3467 | 2140 | 0.3597 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3483 | 2150 | 0.3793 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3500 | 2160 | 0.3633 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3516 | 2170 | 0.3504 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3532 | 2180 | 0.3475 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3548 | 2190 | 0.3467 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3564 | 2200 | 0.3412 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3581 | 2210 | 0.3665 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3597 | 2220 | 0.3585 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3613 | 2230 | 0.3335 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3629 | 2240 | 0.329 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3645 | 2250 | 0.3193 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3662 | 2260 | 0.3256 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3678 | 2270 | 0.325 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3694 | 2280 | 0.3312 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3710 | 2290 | 0.3323 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3726 | 2300 | 0.3192 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3743 | 2310 | 0.3366 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3759 | 2320 | 0.3247 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3775 | 2330 | 0.3207 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3791 | 2340 | 0.3238 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3807 | 2350 | 0.3217 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3824 | 2360 | 0.336 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3840 | 2370 | 0.3043 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3856 | 2380 | 0.3043 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3872 | 2390 | 0.3193 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3888 | 2400 | 0.3145 | 0.2338 | 0.4041 | 0.7329 | 0.2612 | 0.4511 | 0.3624 | 0.2742 | 0.3903 | 0.2020 | 0.2560 | 0.3127 | 0.5038 | 0.4262 | 0.3701 | | 0.3905 | 2410 | 0.319 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3921 | 2420 | 0.3097 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3937 | 2430 | 0.2817 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3953 | 2440 | 0.3168 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3969 | 2450 | 0.2941 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.3986 | 2460 | 0.2902 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4002 | 2470 | 0.3095 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4018 | 2480 | 0.3149 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4034 | 2490 | 0.2949 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4050 | 2500 | 0.3057 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4067 | 2510 | 0.2982 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4083 | 2520 | 0.3064 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4099 | 2530 | 0.3169 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4115 | 2540 | 0.2922 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4131 | 2550 | 0.2999 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4148 | 2560 | 0.2803 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4164 | 2570 | 0.3118 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4180 | 2580 | 0.309 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4196 | 2590 | 0.2894 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4212 | 2600 | 0.3126 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4229 | 2610 | 0.2949 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4245 | 2620 | 0.3204 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4261 | 2630 | 0.2868 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4277 | 2640 | 0.3168 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4293 | 2650 | 0.3245 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4310 | 2660 | 0.316 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4326 | 2670 | 0.2822 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4342 | 2680 | 0.3046 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4358 | 2690 | 0.2908 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4374 | 2700 | 0.2542 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4391 | 2710 | 0.3079 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4407 | 2720 | 0.2821 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4423 | 2730 | 0.2863 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4439 | 2740 | 0.2889 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4455 | 2750 | 0.282 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4472 | 2760 | 0.29 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4488 | 2770 | 0.2973 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4504 | 2780 | 0.3018 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4520 | 2790 | 0.2938 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4536 | 2800 | 0.2835 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4553 | 2810 | 0.2773 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4569 | 2820 | 0.2867 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4585 | 2830 | 0.2954 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4601 | 2840 | 0.3035 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4617 | 2850 | 0.2905 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4634 | 2860 | 0.2821 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4650 | 2870 | 0.2815 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4666 | 2880 | 0.298 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4682 | 2890 | 0.2905 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4698 | 2900 | 0.2821 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4715 | 2910 | 0.2904 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4731 | 2920 | 0.2992 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4747 | 2930 | 0.2834 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4763 | 2940 | 0.2855 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4779 | 2950 | 0.2775 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4796 | 2960 | 0.2994 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4812 | 2970 | 0.2939 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4828 | 2980 | 0.2999 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4844 | 2990 | 0.2935 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4860 | 3000 | 0.2714 | 0.2471 | 0.3962 | 0.7912 | 0.2469 | 0.4488 | 0.3739 | 0.2677 | 0.3976 | 0.1890 | 0.2485 | 0.2962 | 0.4538 | 0.4259 | 0.3679 | | 0.4877 | 3010 | 0.2819 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4893 | 3020 | 0.2679 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4909 | 3030 | 0.2789 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4925 | 3040 | 0.2865 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4941 | 3050 | 0.2852 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4958 | 3060 | 0.2706 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4974 | 3070 | 0.2935 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.4990 | 3080 | 0.272 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5006 | 3090 | 0.2915 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5022 | 3100 | 0.2826 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5039 | 3110 | 0.2652 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5055 | 3120 | 0.2887 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5071 | 3130 | 0.2613 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5087 | 3140 | 0.283 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5103 | 3150 | 0.2945 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5120 | 3160 | 0.2877 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5136 | 3170 | 0.2889 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5152 | 3180 | 0.268 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5168 | 3190 | 0.2911 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5184 | 3200 | 0.2785 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5201 | 3210 | 0.2711 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5217 | 3220 | 0.2911 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5233 | 3230 | 0.2649 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5249 | 3240 | 0.3054 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5265 | 3250 | 0.2531 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5282 | 3260 | 0.2767 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5298 | 3270 | 0.2853 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5314 | 3280 | 0.2731 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5330 | 3290 | 0.2776 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5346 | 3300 | 0.2725 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5363 | 3310 | 0.281 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5379 | 3320 | 0.2666 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5395 | 3330 | 0.2654 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5411 | 3340 | 0.2909 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5427 | 3350 | 0.2598 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5444 | 3360 | 0.2837 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5460 | 3370 | 0.2855 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5476 | 3380 | 0.2601 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5492 | 3390 | 0.268 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5508 | 3400 | 0.2681 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5525 | 3410 | 0.2663 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5541 | 3420 | 0.2837 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5557 | 3430 | 0.259 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5573 | 3440 | 0.2622 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5590 | 3450 | 0.2825 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5606 | 3460 | 0.2921 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5622 | 3470 | 0.2721 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5638 | 3480 | 0.2797 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5654 | 3490 | 0.2899 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5671 | 3500 | 0.2745 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5687 | 3510 | 0.2665 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5703 | 3520 | 0.2908 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5719 | 3530 | 0.2492 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5735 | 3540 | 0.2562 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5752 | 3550 | 0.2616 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5768 | 3560 | 0.2775 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5784 | 3570 | 0.2736 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5800 | 3580 | 0.2862 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5816 | 3590 | 0.2582 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5833 | 3600 | 0.2547 | 0.2371 | 0.3994 | 0.7786 | 0.2418 | 0.4072 | 0.3469 | 0.2615 | 0.4070 | 0.1551 | 0.2294 | 0.2533 | 0.4270 | 0.4161 | 0.3508 | | 0.5849 | 3610 | 0.2822 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5865 | 3620 | 0.2622 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5881 | 3630 | 0.2691 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5897 | 3640 | 0.2585 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5914 | 3650 | 0.2927 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5930 | 3660 | 0.2593 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5946 | 3670 | 0.2501 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5962 | 3680 | 0.2796 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5978 | 3690 | 0.2622 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.5995 | 3700 | 0.2508 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6011 | 3710 | 0.2891 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6027 | 3720 | 0.274 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6043 | 3730 | 0.2769 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6059 | 3740 | 0.2617 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6076 | 3750 | 0.2557 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6092 | 3760 | 0.2634 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6108 | 3770 | 0.262 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6124 | 3780 | 0.2696 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6140 | 3790 | 0.2608 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6157 | 3800 | 0.2592 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6173 | 3810 | 0.2757 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6189 | 3820 | 0.2672 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6205 | 3830 | 0.2523 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6221 | 3840 | 0.2775 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6238 | 3850 | 0.2621 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6254 | 3860 | 0.275 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6270 | 3870 | 0.2727 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6286 | 3880 | 0.2709 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6302 | 3890 | 0.2749 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6319 | 3900 | 0.2844 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6335 | 3910 | 0.2713 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6351 | 3920 | 0.2711 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6367 | 3930 | 0.2523 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6383 | 3940 | 0.2789 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6400 | 3950 | 0.2639 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6416 | 3960 | 0.2609 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6432 | 3970 | 0.2699 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6448 | 3980 | 0.2614 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6464 | 3990 | 0.2567 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6481 | 4000 | 1.2987 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6497 | 4010 | 1.4783 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6513 | 4020 | 1.7162 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6529 | 4030 | 1.2907 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6545 | 4040 | 1.2583 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6562 | 4050 | 1.0498 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6578 | 4060 | 1.8076 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6594 | 4070 | 1.215 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6610 | 4080 | 1.1462 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6626 | 4090 | 0.9511 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6643 | 4100 | 0.6151 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6659 | 4110 | 0.7482 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6675 | 4120 | 0.8572 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6691 | 4130 | 0.7722 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6707 | 4140 | 0.6085 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6724 | 4150 | 0.6644 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6740 | 4160 | 0.6423 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6756 | 4170 | 0.7482 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6772 | 4180 | 0.9649 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6788 | 4190 | 0.9205 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6805 | 4200 | 0.7746 | 0.2822 | 0.4484 | 0.7622 | 0.2944 | 0.5133 | 0.4592 | 0.2717 | 0.4451 | 0.3682 | 0.2594 | 0.2342 | 0.5123 | 0.4209 | 0.4055 | | 0.6821 | 4210 | 0.5752 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6837 | 4220 | 0.6221 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6853 | 4230 | 0.526 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6869 | 4240 | 0.455 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6886 | 4250 | 0.4964 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6902 | 4260 | 0.935 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6918 | 4270 | 0.6227 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6934 | 4280 | 0.5594 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6950 | 4290 | 0.496 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6967 | 4300 | 0.5907 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6983 | 4310 | 0.5163 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.6999 | 4320 | 0.468 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7015 | 4330 | 0.5214 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7031 | 4340 | 0.625 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7048 | 4350 | 0.593 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7064 | 4360 | 0.5852 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7080 | 4370 | 0.5648 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7096 | 4380 | 0.6791 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7112 | 4390 | 0.7008 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7129 | 4400 | 0.6731 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7145 | 4410 | 0.654 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7161 | 4420 | 0.6135 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7177 | 4430 | 0.6206 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7193 | 4440 | 0.5056 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7210 | 4450 | 0.5201 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7226 | 4460 | 0.5894 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7242 | 4470 | 0.5571 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7258 | 4480 | 0.5979 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7274 | 4490 | 0.6202 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7291 | 4500 | 0.5544 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7307 | 4510 | 0.6122 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7323 | 4520 | 0.5631 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7339 | 4530 | 0.5284 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7355 | 4540 | 0.6899 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7372 | 4550 | 0.5838 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7388 | 4560 | 0.6806 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7404 | 4570 | 0.5413 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7420 | 4580 | 0.5956 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7436 | 4590 | 0.6044 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7453 | 4600 | 0.5857 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7469 | 4610 | 0.5664 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7485 | 4620 | 0.5097 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7501 | 4630 | 0.4912 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7517 | 4640 | 0.6049 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7534 | 4650 | 0.5389 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7550 | 4660 | 0.555 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7566 | 4670 | 0.6238 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7582 | 4680 | 0.6447 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7598 | 4690 | 0.5606 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7615 | 4700 | 0.5165 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7631 | 4710 | 0.5839 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7647 | 4720 | 0.5189 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7663 | 4730 | 0.584 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7679 | 4740 | 0.5744 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7696 | 4750 | 0.5351 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7712 | 4760 | 0.5953 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7728 | 4770 | 0.5725 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7744 | 4780 | 0.5688 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7761 | 4790 | 0.5004 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7777 | 4800 | 0.5378 | 0.3514 | 0.4652 | 0.7429 | 0.3103 | 0.5406 | 0.4361 | 0.2797 | 0.4267 | 0.3843 | 0.2727 | 0.3474 | 0.5341 | 0.4249 | 0.4243 | | 0.7793 | 4810 | 0.5244 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7809 | 4820 | 0.6241 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7825 | 4830 | 0.4844 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7842 | 4840 | 0.4401 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7858 | 4850 | 0.499 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7874 | 4860 | 0.5326 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7890 | 4870 | 0.4981 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7906 | 4880 | 0.5659 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7923 | 4890 | 0.5364 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7939 | 4900 | 0.5479 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7955 | 4910 | 0.4653 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7971 | 4920 | 0.5005 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.7987 | 4930 | 0.5624 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8004 | 4940 | 0.4399 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8020 | 4950 | 0.4859 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8036 | 4960 | 0.5087 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8052 | 4970 | 0.511 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8068 | 4980 | 0.5819 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8085 | 4990 | 0.4462 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8101 | 5000 | 0.4882 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8117 | 5010 | 0.5306 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8133 | 5020 | 0.507 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8149 | 5030 | 0.4471 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8166 | 5040 | 0.5333 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8182 | 5050 | 0.4353 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8198 | 5060 | 0.5615 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8214 | 5070 | 0.5629 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8230 | 5080 | 0.5131 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8247 | 5090 | 0.4789 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8263 | 5100 | 0.4934 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8279 | 5110 | 0.5285 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8295 | 5120 | 0.4414 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8311 | 5130 | 0.5262 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8328 | 5140 | 0.4645 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8344 | 5150 | 0.4532 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8360 | 5160 | 0.4421 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8376 | 5170 | 0.4375 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8392 | 5180 | 0.5234 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8409 | 5190 | 0.4803 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8425 | 5200 | 0.4872 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8441 | 5210 | 0.451 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8457 | 5220 | 0.4388 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8473 | 5230 | 0.5182 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8490 | 5240 | 0.5302 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8506 | 5250 | 0.4643 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8522 | 5260 | 0.5581 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8538 | 5270 | 0.4643 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8554 | 5280 | 0.5288 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8571 | 5290 | 0.4133 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8587 | 5300 | 0.4664 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8603 | 5310 | 0.4814 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8619 | 5320 | 0.5256 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8635 | 5330 | 0.4904 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8652 | 5340 | 0.4495 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8668 | 5350 | 0.5389 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8684 | 5360 | 0.4497 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8700 | 5370 | 0.4776 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8716 | 5380 | 0.5441 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8733 | 5390 | 0.4473 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8749 | 5400 | 0.5598 | 0.3381 | 0.4668 | 0.7306 | 0.3137 | 0.5415 | 0.4550 | 0.2840 | 0.4169 | 0.4719 | 0.2735 | 0.3582 | 0.5311 | 0.4076 | 0.4299 | | 0.8765 | 5410 | 0.4726 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8781 | 5420 | 0.4966 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8797 | 5430 | 0.4644 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8814 | 5440 | 0.4084 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8830 | 5450 | 0.4913 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8846 | 5460 | 0.5708 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8862 | 5470 | 0.5577 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8878 | 5480 | 0.4839 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8895 | 5490 | 0.461 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8911 | 5500 | 0.4799 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8927 | 5510 | 0.5608 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8943 | 5520 | 0.4625 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8959 | 5530 | 0.4765 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8976 | 5540 | 0.4348 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.8992 | 5550 | 0.4424 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9008 | 5560 | 0.4147 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9024 | 5570 | 0.433 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9040 | 5580 | 0.4628 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9057 | 5590 | 0.4466 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9073 | 5600 | 0.4563 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9089 | 5610 | 0.4508 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9105 | 5620 | 0.4619 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9121 | 5630 | 0.4264 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9138 | 5640 | 0.5157 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9154 | 5650 | 0.4721 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9170 | 5660 | 0.4518 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9186 | 5670 | 0.4101 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9202 | 5680 | 0.4092 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9219 | 5690 | 0.4042 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9235 | 5700 | 0.3852 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9251 | 5710 | 0.375 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9267 | 5720 | 0.3548 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9283 | 5730 | 0.3461 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9300 | 5740 | 0.3396 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9316 | 5750 | 0.3465 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9332 | 5760 | 0.347 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9348 | 5770 | 0.3365 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9364 | 5780 | 0.3299 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9381 | 5790 | 0.3417 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9397 | 5800 | 0.3423 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9413 | 5810 | 0.3512 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9429 | 5820 | 0.3353 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9445 | 5830 | 0.3291 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9462 | 5840 | 0.3162 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9478 | 5850 | 0.3326 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9494 | 5860 | 0.345 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9510 | 5870 | 0.2998 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9526 | 5880 | 0.307 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9543 | 5890 | 0.3019 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9559 | 5900 | 0.3169 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9575 | 5910 | 0.2857 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9591 | 5920 | 0.3018 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9607 | 5930 | 0.2954 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9624 | 5940 | 0.2953 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9640 | 5950 | 0.2861 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9656 | 5960 | 0.3384 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9672 | 5970 | 0.2968 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9688 | 5980 | 0.3191 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9705 | 5990 | 0.3069 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9721 | 6000 | 0.3025 | 0.3322 | 0.4606 | 0.6623 | 0.3084 | 0.5552 | 0.4463 | 0.2714 | 0.4404 | 0.7084 | 0.2888 | 0.3529 | 0.4924 | 0.4138 | 0.4410 | | 0.9737 | 6010 | 0.2891 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9753 | 6020 | 0.3038 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9769 | 6030 | 0.2931 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9786 | 6040 | 0.3145 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9802 | 6050 | 0.3046 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9818 | 6060 | 0.2896 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9834 | 6070 | 0.2926 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9850 | 6080 | 0.3025 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9867 | 6090 | 0.2798 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9883 | 6100 | 0.3006 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9899 | 6110 | 0.2695 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9915 | 6120 | 0.3017 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9931 | 6130 | 0.2955 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9948 | 6140 | 0.2699 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9964 | 6150 | 0.2955 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9980 | 6160 | 0.2963 | - | - | - | - | - | - | - | - | - | - | - | - | - | - | | 0.9996 | 6170 | 0.2988 | - | - | - | - | - | - | - | - | - | - | - | - | - | - |
### Framework Versions - Python: 3.10.12 - Sentence Transformers: 3.3.0 - Transformers: 4.46.2 - PyTorch: 2.1.1+cu121 - Accelerate: 0.34.2 - Datasets: 3.1.0 - Tokenizers: 0.20.3 ## Citation ### BibTeX #### Sentence Transformers ```bibtex @inproceedings{reimers-2019-sentence-bert, title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks", author = "Reimers, Nils and Gurevych, Iryna", booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing", month = "11", year = "2019", publisher = "Association for Computational Linguistics", url = "https://arxiv.org/abs/1908.10084", } ``` #### MultipleNegativesRankingLoss ```bibtex @misc{henderson2017efficient, title={Efficient Natural Language Response Suggestion for Smart Reply}, author={Matthew Henderson and Rami Al-Rfou and Brian Strope and Yun-hsuan Sung and Laszlo Lukacs and Ruiqi Guo and Sanjiv Kumar and Balint Miklos and Ray Kurzweil}, year={2017}, eprint={1705.00652}, archivePrefix={arXiv}, primaryClass={cs.CL} } ```