Ananthu357's picture
Add new SentenceTransformer model.
9b4c43d verified
metadata
base_model: BAAI/bge-large-en
datasets: []
language: []
library_name: sentence-transformers
pipeline_tag: sentence-similarity
tags:
  - sentence-transformers
  - sentence-similarity
  - feature-extraction
  - generated_from_trainer
  - dataset_size:649
  - loss:CosineSimilarityLoss
widget:
  - source_sentence: Dispute resolution
    sentences:
      - >-
        Arbitration and Conciliation (Amendment) Act 2015, if they agree for
        such waiver in writing, after dispute having arisen between them, in the
        format
      - >-
        The Earnest Money shall be deposited in cash through e-payment gateway
        or as mentioned in tender documents.
      - >-
        of liquidated damages under this condition shall not exceed 5% of the
        contract value
  - source_sentence: >-
      Order of Precedence is the order with which preference should be given to
      the documents.
    sentences:
      - the sand, stone, clay ballast, earth, trees, rock
      - >-
        in case of any difference, contradiction, discrepancy, with regard to
        conditions of tender/contract,
      - >-
        If the tenderer(s) deliberately gives / give wrong information in his /
        their tender or creates / create circumstances for the acceptance of his
        / their tender, the Railway reserves the right to reject such tender at
        any stage.
  - source_sentence: >-
      Does the contract document contain a 'third-party liability relationship'
      provision?
    sentences:
      - >-
        The Contractor shall be responsible for all risk to the work and for
        trespass and shall make good at his own expense all loss or damage
        whether to the works themselves or to any other property of the Railway
        or the lives, persons or property of other
      - >-
        This program should indicate the time schedule for various work items in
        the form of a Bar Chart/PERT/CPM.
      - "\_Completion indiacted by issuance of maintenance certifciate"
  - source_sentence: What is the impact of breaching the contract conditions on subcontracting?
    sentences:
      - Schedule of Rates
      - What determines the completion of the contract.
      - >-
        shall not assign or sublet the contract or any part thereof or allow any
        person
  - source_sentence: Bonus for early completion of work
    sentences:
      - >-
        as to execution or quality of any work or material, or as to the
        measurements of the works the decision of the Engineer thereon shall be
        final subject to the appeal (within 7 days of such decision being
        intimated to the Contractor) to the Chief Engineer 
      - The maximum bonus shall be limited to 3% of original contract value.
      - >-
        The Contractor shall indemnify and save harmless the Railway from and
        against all actions, suit, proceedings, losses, costs, damages, charges,
        claims and demands of every nature and description brought or recovered
        against the Railways by reason of any act or omission of the Contractor,
        his agents or employees, in the execution of the works or in his
        guarding of the same. All sums payable by way of compensation under any
        of these conditions shall be considered as reasonable compensation to be
        applied to the actual loss or damage sustained, and whether or not any
        damage shall have been sustained.

SentenceTransformer based on BAAI/bge-large-en

This is a sentence-transformers model finetuned from BAAI/bge-large-en. It maps sentences & paragraphs to a 1024-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.

Model Details

Model Description

  • Model Type: Sentence Transformer
  • Base model: BAAI/bge-large-en
  • Maximum Sequence Length: 512 tokens
  • Output Dimensionality: 1024 tokens
  • Similarity Function: Cosine Similarity

Model Sources

Full Model Architecture

SentenceTransformer(
  (0): Transformer({'max_seq_length': 512, 'do_lower_case': True}) with Transformer model: BertModel 
  (1): Pooling({'word_embedding_dimension': 1024, 'pooling_mode_cls_token': True, 'pooling_mode_mean_tokens': False, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
  (2): Normalize()
)

Usage

Direct Usage (Sentence Transformers)

First install the Sentence Transformers library:

pip install -U sentence-transformers

Then you can load this model and run inference.

from sentence_transformers import SentenceTransformer

# Download from the 🤗 Hub
model = SentenceTransformer("Ananthu357/Ananthus-BAAI-for-contracts11.0")
# Run inference
sentences = [
    'Bonus for early completion of work',
    'The maximum bonus shall be limited to 3% of original contract value.',
    'The Contractor shall indemnify and save harmless the Railway from and against all actions, suit, proceedings, losses, costs, damages, charges, claims and demands of every nature and description brought or recovered against the Railways by reason of any act or omission of the Contractor, his agents or employees, in the execution of the works or in his guarding of the same. All sums payable by way of compensation under any of these conditions shall be considered as reasonable compensation to be applied to the actual loss or damage sustained, and whether or not any damage shall have been sustained.',
]
embeddings = model.encode(sentences)
print(embeddings.shape)
# [3, 1024]

# Get the similarity scores for the embeddings
similarities = model.similarity(embeddings, embeddings)
print(similarities.shape)
# [3, 3]

Training Details

Training Hyperparameters

Non-Default Hyperparameters

  • eval_strategy: steps
  • per_device_train_batch_size: 16
  • per_device_eval_batch_size: 16
  • num_train_epochs: 15
  • warmup_ratio: 0.1
  • fp16: True
  • batch_sampler: no_duplicates

All Hyperparameters

Click to expand
  • overwrite_output_dir: False
  • do_predict: False
  • eval_strategy: steps
  • prediction_loss_only: True
  • per_device_train_batch_size: 16
  • per_device_eval_batch_size: 16
  • per_gpu_train_batch_size: None
  • per_gpu_eval_batch_size: None
  • gradient_accumulation_steps: 1
  • eval_accumulation_steps: None
  • learning_rate: 5e-05
  • weight_decay: 0.0
  • adam_beta1: 0.9
  • adam_beta2: 0.999
  • adam_epsilon: 1e-08
  • max_grad_norm: 1.0
  • num_train_epochs: 15
  • max_steps: -1
  • lr_scheduler_type: linear
  • lr_scheduler_kwargs: {}
  • warmup_ratio: 0.1
  • warmup_steps: 0
  • log_level: passive
  • log_level_replica: warning
  • log_on_each_node: True
  • logging_nan_inf_filter: True
  • save_safetensors: True
  • save_on_each_node: False
  • save_only_model: False
  • restore_callback_states_from_checkpoint: False
  • no_cuda: False
  • use_cpu: False
  • use_mps_device: False
  • seed: 42
  • data_seed: None
  • jit_mode_eval: False
  • use_ipex: False
  • bf16: False
  • fp16: True
  • fp16_opt_level: O1
  • half_precision_backend: auto
  • bf16_full_eval: False
  • fp16_full_eval: False
  • tf32: None
  • local_rank: 0
  • ddp_backend: None
  • tpu_num_cores: None
  • tpu_metrics_debug: False
  • debug: []
  • dataloader_drop_last: False
  • dataloader_num_workers: 0
  • dataloader_prefetch_factor: None
  • past_index: -1
  • disable_tqdm: False
  • remove_unused_columns: True
  • label_names: None
  • load_best_model_at_end: False
  • ignore_data_skip: False
  • fsdp: []
  • fsdp_min_num_params: 0
  • fsdp_config: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}
  • fsdp_transformer_layer_cls_to_wrap: None
  • accelerator_config: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}
  • deepspeed: None
  • label_smoothing_factor: 0.0
  • optim: adamw_torch
  • optim_args: None
  • adafactor: False
  • group_by_length: False
  • length_column_name: length
  • ddp_find_unused_parameters: None
  • ddp_bucket_cap_mb: None
  • ddp_broadcast_buffers: False
  • dataloader_pin_memory: True
  • dataloader_persistent_workers: False
  • skip_memory_metrics: True
  • use_legacy_prediction_loop: False
  • push_to_hub: False
  • resume_from_checkpoint: None
  • hub_model_id: None
  • hub_strategy: every_save
  • hub_private_repo: False
  • hub_always_push: False
  • gradient_checkpointing: False
  • gradient_checkpointing_kwargs: None
  • include_inputs_for_metrics: False
  • eval_do_concat_batches: True
  • fp16_backend: auto
  • push_to_hub_model_id: None
  • push_to_hub_organization: None
  • mp_parameters:
  • auto_find_batch_size: False
  • full_determinism: False
  • torchdynamo: None
  • ray_scope: last
  • ddp_timeout: 1800
  • torch_compile: False
  • torch_compile_backend: None
  • torch_compile_mode: None
  • dispatch_batches: None
  • split_batches: None
  • include_tokens_per_second: False
  • include_num_input_tokens_seen: False
  • neftune_noise_alpha: None
  • optim_target_modules: None
  • batch_eval_metrics: False
  • eval_on_start: False
  • batch_sampler: no_duplicates
  • multi_dataset_batch_sampler: proportional

Training Logs

Epoch Step Training Loss loss
2.4390 100 0.0672 0.0435
4.8780 200 0.0132 0.0396
7.3171 300 0.0052 0.0404
9.7561 400 0.0027 0.0419
12.1951 500 0.002 0.0420
14.6341 600 0.0014 0.0423

Framework Versions

  • Python: 3.10.12
  • Sentence Transformers: 3.0.1
  • Transformers: 4.42.4
  • PyTorch: 2.3.1+cu121
  • Accelerate: 0.32.1
  • Datasets: 2.21.0
  • Tokenizers: 0.19.1

Citation

BibTeX

Sentence Transformers

@inproceedings{reimers-2019-sentence-bert,
    title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
    author = "Reimers, Nils and Gurevych, Iryna",
    booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
    month = "11",
    year = "2019",
    publisher = "Association for Computational Linguistics",
    url = "https://arxiv.org/abs/1908.10084",
}