From 4d0d9cbaaaf2d548d7c42c7e4ec546c0d85e7143 Mon Sep 17 00:00:00 2001 From: Manodeepray Date: Fri, 13 Jun 2025 02:06:50 +0530 Subject: [PATCH 1/5] Fix: Resolve issue #12 --- finetune.py | 196 ++++++++++++++++++++++++++++++++ peft_configs/lora_configs.yaml | 11 ++ peft_configs/qlora_configs.yaml | 15 +++ scripts/run_lora.sh | 12 ++ scripts/run_qlora.sh | 10 ++ scripts/run_training.sh | 9 ++ train.py | 26 ++++- 7 files changed, 276 insertions(+), 3 deletions(-) create mode 100644 finetune.py create mode 100644 peft_configs/lora_configs.yaml create mode 100644 peft_configs/qlora_configs.yaml create mode 100644 scripts/run_lora.sh create mode 100644 scripts/run_qlora.sh create mode 100644 scripts/run_training.sh diff --git a/finetune.py b/finetune.py new file mode 100644 index 0000000..5f7b4d1 --- /dev/null +++ b/finetune.py @@ -0,0 +1,196 @@ +import logging +import wandb +from functools import partial + +import torch +from datasets import load_dataset +from torch.utils.data import DataLoader +from transformers import AutoProcessor, Gemma3ForConditionalGeneration +from peft import LoraConfig, get_peft_model + +from config import Configuration +from utils import train_collate_function +import argparse +import albumentations as A +import yaml +from tqdm import tqdm +from transformers import BitsAndBytesConfig +from peft import prepare_model_for_kbit_training + + + + + + + +logging.basicConfig( + level=logging.INFO, format="%(asctime)s - %(name)s - %(levelname)s - %(message)s" +) +logger = logging.getLogger(__name__) + +augmentations = A.Compose([ + A.Resize(height=896, width=896), + A.HorizontalFlip(p=0.5), + A.ColorJitter(p=0.2), +], bbox_params=A.BboxParams(format='coco', label_fields=['category_ids'], filter_invalid_bboxes=True)) + +def get_dataloader(processor, args, dtype): + logger.info("Fetching the dataset") + train_dataset = load_dataset(cfg.dataset_id, split="train") + train_collate_fn = partial( + train_collate_function, processor=processor, dtype=dtype, transform=augmentations + ) + + logger.info("Building data loader") + train_dataloader = DataLoader( + train_dataset, + batch_size=args.batch_size, + collate_fn=train_collate_fn, + shuffle=True, + ) + return train_dataloader + +def train_model(model, optimizer, cfg, train_dataloader): + logger.info("Start training") + global_step = 0 + + + epoch_pbar = tqdm(range(cfg.epochs) , desc = "Epochs" , position= 0) + + + for epoch in epoch_pbar: + + epoch_pbar.set_description(f"Epoch {epoch+1}/{cfg.epochs}") + + + + batch_pbar = tqdm(train_dataloader, desc="Batches", leave=False, position=1) + + for idx, batch in enumerate(batch_pbar): + outputs = model(**batch.to(model.device)) + loss = outputs.loss + if idx % 100 == 0: + logger.info(f"Epoch: {epoch} Iter: {idx} Loss: {loss.item():.4f}") + wandb.log({"train/loss": loss.item(), "epoch": epoch}, step=global_step) + + loss.backward() + optimizer.step() + optimizer.zero_grad() + global_step += 1 + batch_pbar.close() + + epoch_pbar.close() + return model + + +def get_peft_config(peft_type: str, config_dict: dict) -> LoraConfig: + """Factory method to create PEFT config based on type""" + common_config = { + "r": config_dict["r"], + "lora_alpha": config_dict["lora_alpha"], + "target_modules": config_dict["target_modules"], + "lora_dropout": config_dict["lora_dropout"], + "bias": config_dict["bias"], + "task_type": config_dict["task_type"], + } + + if peft_type == "qlora": + # Add QLoRA specific configurations if needed + common_config.update({ + "use_dora": config_dict.get("use_dora", False), # DORA: Weight-Decomposed Low-Rank Adaptation + }) + return LoraConfig(**common_config) + + + +if __name__ == "__main__": + cfg = Configuration.from_args() + + # Get values dynamically from user + parser = argparse.ArgumentParser(description="Training for PaLiGemma") + parser.add_argument('--model_id', type=str, required=True, default=cfg.model_id, help='Enter Huggingface Model ID') + parser.add_argument('--dataset_id', type=str, required=True ,default=cfg.dataset_id, help='Enter Huggingface Dataset ID') + parser.add_argument('--batch_size', type=int, default=cfg.batch_size, help='Enter Batch Size') + parser.add_argument('--lr', type=float, default=cfg.learning_rate, help='Enter Learning Rate') + parser.add_argument('--checkpoint_id', type=str, required=True, default=cfg.checkpoint_id, help='Enter Huggingface Repo ID to push model') + + + parser.add_argument('--peft_type', type=str, required=True, choices=["lora" , "qlora"] ,help='Enter peft type .for eg. lora , qlora ..etc') + parser.add_argument('--peft_config', type=str, default="peft_configs/lora_configs.yaml", + help="Path to peft config YAML file") + + + + + + args = parser.parse_args() + processor = AutoProcessor.from_pretrained(args.model_id) + train_dataloader = get_dataloader(processor=processor, args=args, dtype=cfg.dtype) + + logger.info("Getting model") + + + bnb_config = BitsAndBytesConfig( + load_in_4bit=True, + bnb_4bit_use_double_quant=True, + bnb_4bit_quant_type="nf4", + bnb_4bit_compute_dtype=cfg.dtype, + ) + + + + model = Gemma3ForConditionalGeneration.from_pretrained( + cfg.model_id, + torch_dtype=cfg.dtype, + device_map="cpu", + attn_implementation="eager", + quantization_config=bnb_config if args.peft_type == "qlora" else None + ) + logger.info(f"Loading LoRA config from {args.peft_type}") + + + + + if args.peft_type == "lora": + with open(args.lora_config) as f: + lora_config_dict = yaml.safe_load(f)[f"{args.peft_type}config"] + + lora_config = get_peft_config(peft_type=args.peft_type , config_dict=lora_config_dict) + + model = get_peft_model(model, lora_config) + model.print_trainable_parameters() + + + if args.peft_type =='qlora': + + + with open(args.peft_config) as f: + qlora_config_dict = yaml.safe_load(f)[f"{args.peft_type}_config"] + + peft_config = get_peft_config(args.peft_type, qlora_config_dict) + + model = prepare_model_for_kbit_training(model) + model = get_peft_model(model, peft_config) + model.print_trainable_parameters() + + + model.train() + model.to(cfg.device) + + params_to_train = list(filter(lambda x: x.requires_grad, model.parameters())) + optimizer = torch.optim.AdamW(params_to_train, lr=args.lr) + + wandb.init( + project=cfg.project_name, + name=cfg.run_name if hasattr(cfg, "run_name") else None, + config=vars(cfg), + ) + + train_model(model, optimizer, cfg, train_dataloader) + + # Push the checkpoint to hub + model.push_to_hub(cfg.checkpoint_id) + processor.push_to_hub(cfg.checkpoint_id) + + wandb.finish() + logger.info("Train finished") \ No newline at end of file diff --git a/peft_configs/lora_configs.yaml b/peft_configs/lora_configs.yaml new file mode 100644 index 0000000..a993975 --- /dev/null +++ b/peft_configs/lora_configs.yaml @@ -0,0 +1,11 @@ +lora_config: + r: 8 # Rank + lora_alpha: 16 # Scaling factor + target_modules: # Modules to apply LoRA to + - "q_proj" + - "k_proj" + - "v_proj" + - "o_proj" + lora_dropout: 0.05 # Dropout probability + bias: "none" # Bias type + task_type: "CAUSAL_LM" \ No newline at end of file diff --git a/peft_configs/qlora_configs.yaml b/peft_configs/qlora_configs.yaml new file mode 100644 index 0000000..073a484 --- /dev/null +++ b/peft_configs/qlora_configs.yaml @@ -0,0 +1,15 @@ +qlora_config: + r: 64 # Higher rank often works better with QLoRA + lora_alpha: 16 + target_modules: + - "q_proj" + - "k_proj" + - "v_proj" + - "o_proj" + - "gate_proj" # Additional modules for better performance + - "up_proj" + - "down_proj" + lora_dropout: 0.1 # Slightly higher dropout + bias: "none" + task_type: "CAUSAL_LM" + use_dora: true # Enable DORA for better performance \ No newline at end of file diff --git a/scripts/run_lora.sh b/scripts/run_lora.sh new file mode 100644 index 0000000..6540405 --- /dev/null +++ b/scripts/run_lora.sh @@ -0,0 +1,12 @@ + +!#bin/bash + + +python finetune.py \ + --model_id google/gemma-3-4b-pt \ + --dataset_id ariG23498/license-detection-paligemma \ + --batch_size 8 \ + --lr 2e-5 \ + --checkpoint_id oreonmayo/gemma-3-4b-pt-object-detection-aug \ + --peft_type lora \ + --peft_config peft_configs/lora_configs.yaml \ \ No newline at end of file diff --git a/scripts/run_qlora.sh b/scripts/run_qlora.sh new file mode 100644 index 0000000..286a690 --- /dev/null +++ b/scripts/run_qlora.sh @@ -0,0 +1,10 @@ +!#bin/bash + +python finetune.py \ + --model_id google/gemma-3-4b-pt \ + --dataset_id ariG23498/license-detection-paligemma \ + --batch_size 8 \ + --lr 2e-5 \ + --checkpoint_id oreonmayo/gemma-3-4b-pt-object-detection-aug \ + --peft_type qlora \ + --peft_config peft_configs/qlora_configs.yaml \ \ No newline at end of file diff --git a/scripts/run_training.sh b/scripts/run_training.sh new file mode 100644 index 0000000..4cd3f16 --- /dev/null +++ b/scripts/run_training.sh @@ -0,0 +1,9 @@ +!#bin/bash + + +python train.py \ + --model_id google/gemma-3-4b-pt \ + --dataset_id ariG23498/license-detection-paligemma \ + --batch_size 8 \ + --lr 2e-5 \ + --checkpoint_id oreonmayo/gemma-3-4b-pt-object-detection-aug \ No newline at end of file diff --git a/train.py b/train.py index 8aab73d..7989c4c 100644 --- a/train.py +++ b/train.py @@ -12,6 +12,11 @@ import argparse import albumentations as A +from tqdm import tqdm + + + + logging.basicConfig( level=logging.INFO, format="%(asctime)s - %(name)s - %(levelname)s - %(message)s" ) @@ -45,8 +50,20 @@ def get_dataloader(processor, args, dtype): def train_model(model, optimizer, cfg, train_dataloader): logger.info("Start training") global_step = 0 - for epoch in range(cfg.epochs): - for idx, batch in enumerate(train_dataloader): + + + epoch_pbar = tqdm(range(cfg.epochs) , desc = "Epochs" , position= 0) + + + for epoch in epoch_pbar: + + epoch_pbar.set_description(f"Epoch {epoch+1}/{cfg.epochs}") + + + + batch_pbar = tqdm(train_dataloader, desc="Batches", leave=False, position=1) + + for idx, batch in enumerate(batch_pbar): outputs = model(**batch.to(model.device)) loss = outputs.loss if idx % 100 == 0: @@ -57,11 +74,14 @@ def train_model(model, optimizer, cfg, train_dataloader): optimizer.step() optimizer.zero_grad() global_step += 1 + batch_pbar.close() + + epoch_pbar.close() return model if __name__ == "__main__": - cfg = Configuration.from_args() + cfg = Configuration() # Get values dynamicaly from user parser = argparse.ArgumentParser(description="Training for PaLiGemma") From 34bbcc1898952f8e7e0f600a2417e55c57a6c564 Mon Sep 17 00:00:00 2001 From: Manodeepray Date: Fri, 13 Jun 2025 02:21:40 +0530 Subject: [PATCH 2/5] Fix: Resolve issue #12 --- config.py | 2 + finetune.py | 25 +- scripts/run_lora.sh | 4 +- scripts/run_qlora.sh | 4 +- scripts/run_training.sh | 2 +- wandb/debug-internal.log | 1 + wandb/debug.log | 1 + wandb/latest-run | 1 + .../files/output.log | 0 .../files/requirements.txt | 253 ++++++++++++++++++ .../files/wandb-metadata.json | 42 +++ .../logs/debug-core.log | 1 + .../logs/debug-internal.log | 8 + .../logs/debug.log | 21 ++ .../run-6ot3basl.wandb | 0 .../files/output.log | 0 .../files/requirements.txt | 253 ++++++++++++++++++ .../files/wandb-metadata.json | 46 ++++ .../logs/debug-core.log | 1 + .../logs/debug-internal.log | 8 + .../logs/debug.log | 21 ++ .../run-ok4hunfv.wandb | 0 22 files changed, 677 insertions(+), 17 deletions(-) create mode 120000 wandb/debug-internal.log create mode 120000 wandb/debug.log create mode 120000 wandb/latest-run create mode 100644 wandb/offline-run-20250613_021623-6ot3basl/files/output.log create mode 100644 wandb/offline-run-20250613_021623-6ot3basl/files/requirements.txt create mode 100644 wandb/offline-run-20250613_021623-6ot3basl/files/wandb-metadata.json create mode 120000 wandb/offline-run-20250613_021623-6ot3basl/logs/debug-core.log create mode 100644 wandb/offline-run-20250613_021623-6ot3basl/logs/debug-internal.log create mode 100644 wandb/offline-run-20250613_021623-6ot3basl/logs/debug.log create mode 100644 wandb/offline-run-20250613_021623-6ot3basl/run-6ot3basl.wandb create mode 100644 wandb/offline-run-20250613_022053-ok4hunfv/files/output.log create mode 100644 wandb/offline-run-20250613_022053-ok4hunfv/files/requirements.txt create mode 100644 wandb/offline-run-20250613_022053-ok4hunfv/files/wandb-metadata.json create mode 120000 wandb/offline-run-20250613_022053-ok4hunfv/logs/debug-core.log create mode 100644 wandb/offline-run-20250613_022053-ok4hunfv/logs/debug-internal.log create mode 100644 wandb/offline-run-20250613_022053-ok4hunfv/logs/debug.log create mode 100644 wandb/offline-run-20250613_022053-ok4hunfv/run-ok4hunfv.wandb diff --git a/config.py b/config.py index 333d864..becebce 100644 --- a/config.py +++ b/config.py @@ -16,4 +16,6 @@ class Configuration: batch_size: int = 8 learning_rate: float = 2e-05 epochs = 2 + + project_name: str = "gemma_finetune" diff --git a/finetune.py b/finetune.py index 5f7b4d1..87fb680 100644 --- a/finetune.py +++ b/finetune.py @@ -104,7 +104,7 @@ def get_peft_config(peft_type: str, config_dict: dict) -> LoraConfig: if __name__ == "__main__": - cfg = Configuration.from_args() + cfg = Configuration() # Get values dynamically from user parser = argparse.ArgumentParser(description="Training for PaLiGemma") @@ -129,14 +129,15 @@ def get_peft_config(peft_type: str, config_dict: dict) -> LoraConfig: logger.info("Getting model") - - bnb_config = BitsAndBytesConfig( - load_in_4bit=True, - bnb_4bit_use_double_quant=True, - bnb_4bit_quant_type="nf4", - bnb_4bit_compute_dtype=cfg.dtype, - ) - + if args.peft_type == 'qlora': + bnb_config = BitsAndBytesConfig( + load_in_4bit=True, + bnb_4bit_use_double_quant=True, + bnb_4bit_quant_type="nf4", + bnb_4bit_compute_dtype=cfg.dtype, + ) + else: + bnb_config = None model = Gemma3ForConditionalGeneration.from_pretrained( @@ -146,14 +147,14 @@ def get_peft_config(peft_type: str, config_dict: dict) -> LoraConfig: attn_implementation="eager", quantization_config=bnb_config if args.peft_type == "qlora" else None ) - logger.info(f"Loading LoRA config from {args.peft_type}") + logger.info(f"Loading PeFT config from {args.peft_type}") if args.peft_type == "lora": - with open(args.lora_config) as f: - lora_config_dict = yaml.safe_load(f)[f"{args.peft_type}config"] + with open(args.peft_config) as f: + lora_config_dict = yaml.safe_load(f)[f"{args.peft_type}_config"] lora_config = get_peft_config(peft_type=args.peft_type , config_dict=lora_config_dict) diff --git a/scripts/run_lora.sh b/scripts/run_lora.sh index 6540405..2c55ab1 100644 --- a/scripts/run_lora.sh +++ b/scripts/run_lora.sh @@ -1,5 +1,5 @@ -!#bin/bash +#!bin/bash python finetune.py \ @@ -9,4 +9,4 @@ python finetune.py \ --lr 2e-5 \ --checkpoint_id oreonmayo/gemma-3-4b-pt-object-detection-aug \ --peft_type lora \ - --peft_config peft_configs/lora_configs.yaml \ \ No newline at end of file + --peft_config peft_configs/lora_configs.yaml \ No newline at end of file diff --git a/scripts/run_qlora.sh b/scripts/run_qlora.sh index 286a690..8585398 100644 --- a/scripts/run_qlora.sh +++ b/scripts/run_qlora.sh @@ -1,4 +1,4 @@ -!#bin/bash +#!bin/bash python finetune.py \ --model_id google/gemma-3-4b-pt \ @@ -7,4 +7,4 @@ python finetune.py \ --lr 2e-5 \ --checkpoint_id oreonmayo/gemma-3-4b-pt-object-detection-aug \ --peft_type qlora \ - --peft_config peft_configs/qlora_configs.yaml \ \ No newline at end of file + --peft_config peft_configs/qlora_configs.yaml \ No newline at end of file diff --git a/scripts/run_training.sh b/scripts/run_training.sh index 4cd3f16..7eef447 100644 --- a/scripts/run_training.sh +++ b/scripts/run_training.sh @@ -1,4 +1,4 @@ -!#bin/bash +#!bin/bash python train.py \ diff --git a/wandb/debug-internal.log b/wandb/debug-internal.log new file mode 120000 index 0000000..6955c19 --- /dev/null +++ b/wandb/debug-internal.log @@ -0,0 +1 @@ +offline-run-20250613_022053-ok4hunfv/logs/debug-internal.log \ No newline at end of file diff --git a/wandb/debug.log b/wandb/debug.log new file mode 120000 index 0000000..6b93e96 --- /dev/null +++ b/wandb/debug.log @@ -0,0 +1 @@ +offline-run-20250613_022053-ok4hunfv/logs/debug.log \ No newline at end of file diff --git a/wandb/latest-run b/wandb/latest-run new file mode 120000 index 0000000..430abf7 --- /dev/null +++ b/wandb/latest-run @@ -0,0 +1 @@ +offline-run-20250613_022053-ok4hunfv \ No newline at end of file diff --git a/wandb/offline-run-20250613_021623-6ot3basl/files/output.log b/wandb/offline-run-20250613_021623-6ot3basl/files/output.log new file mode 100644 index 0000000..e69de29 diff --git a/wandb/offline-run-20250613_021623-6ot3basl/files/requirements.txt b/wandb/offline-run-20250613_021623-6ot3basl/files/requirements.txt new file mode 100644 index 0000000..2b4eca9 --- /dev/null +++ b/wandb/offline-run-20250613_021623-6ot3basl/files/requirements.txt @@ -0,0 +1,253 @@ +Deprecated==1.2.18 +zipp==3.22.0 +langgraph-sdk==0.1.70 +psutil==7.0.0 +arrow==1.3.0 +aiohappyeyeballs==2.6.1 +charset-normalizer==3.4.2 +idna==3.10 +ptyprocess==0.7.0 +jupyter-lsp==2.2.5 +certifi==2025.4.26 +GitPython==3.1.44 +argparse==1.4.0 +webcolors==24.11.1 +bitsandbytes==0.46.0 +pandocfilters==1.5.1 +opentelemetry-sdk==1.33.1 +Flask==3.1.1 +ipython_pygments_lexers==1.1.1 +overrides==7.7.0 +notebook==7.4.2 +sympy==1.14.0 +cffi==1.17.1 +Send2Trash==1.8.3 +seaborn==0.13.2 +jedi==0.19.2 +anyio==4.9.0 +uri-template==1.3.0 +kiwisolver==1.4.8 +importlib_metadata==8.6.1 +pyppeteer==2.0.0 +stack-data==0.6.3 +comm==0.2.2 +tensorboard==2.19.0 +mistune==3.1.3 +isoduration==20.11.0 +nvidia-cusparse-cu12==12.5.4.2 +argon2-cffi-bindings==21.2.0 +accelerate==1.7.0 +wrapt==1.17.2 +stringzilla==3.12.5 +httpcore==1.0.9 +distro==1.9.0 +rpds-py==0.24.0 +python-json-logger==3.3.0 +pyasn1_modules==0.4.2 +requests-toolbelt==1.0.0 +Markdown==3.8 +aiohttp==3.12.6 +rsa==4.9.1 +albucore==0.0.24 +trl==0.18.1 +graphene==3.4.3 +nvidia-cublas-cu12==12.6.4.1 +typing-inspection==0.4.1 +threadpoolctl==3.6.0 +protobuf==6.31.0 +zstandard==0.23.0 +pandas==2.2.3 +langchain-groq==0.3.2 +nbconvert==7.16.6 +multidict==6.4.4 +babel==2.17.0 +opentelemetry-api==1.33.1 +torchaudio==2.7.0 +tzdata==2025.2 +dataclasses-json==0.6.7 +jsonschema==4.23.0 +matplotlib-inline==0.1.7 +networkx==3.5 +setuptools==80.9.0 +mpmath==1.3.0 +alembic==1.16.1 +prometheus_client==0.21.1 +argon2-cffi==23.1.0 +nvidia-cusparselt-cu12==0.6.3 +frozenlist==1.6.0 +pycparser==2.22 +jupyterlab_widgets==3.0.15 +multiprocess==0.70.16 +nvidia-curand-cu12==10.3.7.77 +langchain-community==0.3.24 +xxhash==3.5.0 +websocket-client==1.8.0 +soupsieve==2.7 +gitdb==4.0.12 +notebook_shim==0.2.4 +fonttools==4.57.0 +ipykernel==6.29.5 +graphql-core==3.2.6 +langchain-text-splitters==0.3.8 +langchain-core==0.3.61 +scikit-learn==1.6.1 +ormsgpack==1.9.1 +jupyter_core==5.7.2 +jupyter_server_terminals==0.5.3 +hf-xet==1.1.2 +jsonschema-specifications==2025.4.1 +opentelemetry-semantic-conventions==0.54b1 +nvidia-cuda-runtime-cu12==12.6.77 +tokenizers==0.21.1 +huggingface-hub==0.32.3 +torch==2.7.0 +nbclient==0.10.2 +websockets==10.4 +google-auth==2.40.2 +databricks-sdk==0.54.0 +fsspec==2025.3.0 +setproctitle==1.3.6 +PyYAML==6.0.2 +nvidia-cudnn-cu12==9.5.1.17 +marshmallow==3.26.1 +requests==2.32.3 +pyarrow==20.0.0 +langgraph==0.4.5 +pydantic_core==2.33.2 +langgraph-prebuilt==0.2.1 +prompt_toolkit==3.0.51 +nvidia-nvjitlink-cu12==12.6.85 +appdirs==1.4.4 +tinycss2==1.4.0 +asttokens==3.0.0 +wandb==0.20.1 +antlr4-python3-runtime==4.13.2 +tqdm==4.67.1 +bash_kernel==0.10.0 +nvidia-nccl-cu12==2.26.2 +numpy==2.2.6 +fastapi==0.115.12 +gunicorn==23.0.0 +attrs==25.3.0 +latex2sympy2_extended==1.10.1 +urllib3==2.4.0 +jsonpointer==3.0.0 +matplotlib==3.10.1 +pyee==11.1.1 +nvidia-cuda-cupti-cu12==12.6.80 +markdown-it-py==3.0.0 +httpx==0.28.1 +nest-asyncio==1.6.0 +pyasn1==0.6.1 +datasets==3.6.0 +decorator==5.2.1 +mdurl==0.1.2 +dill==0.3.8 +nvidia-cufft-cu12==11.3.0.4 +jupyter_server==2.15.0 +click==8.2.1 +jsonpatch==1.33 +webencodings==0.5.1 +defusedxml==0.7.1 +simsimd==6.4.9 +terminado==0.18.1 +tiktoken==0.9.0 +graphql-relay==3.2.0 +pexpect==4.9.0 +joblib==1.4.2 +rfc3339-validator==0.1.4 +rich==14.0.0 +tenacity==9.1.2 +widgetsnbextension==4.0.14 +h11==0.16.0 +debugpy==1.8.14 +nvidia-cufile-cu12==1.11.1.6 +filetype==1.2.0 +nvidia-cuda-nvrtc-cu12==12.6.77 +types-python-dateutil==2.9.0.20241206 +albumentations==2.0.8 +peft==0.15.2 +python-dateutil==2.9.0.post0 +rfc3986-validator==0.1.1 +bleach==6.2.0 +Pygments==2.19.1 +sentry-sdk==2.30.0 +propcache==0.3.1 +langsmith==0.3.42 +jupyterlab_pygments==0.3.0 +docker==7.1.0 +mypy_extensions==1.1.0 +groq==0.25.0 +mlflow==2.22.0 +pydantic-settings==2.9.1 +langgraph-checkpoint==2.0.26 +pure_eval==0.2.3 +sqlparse==0.5.3 +httpx-sse==0.4.0 +six==1.17.0 +filelock==3.18.0 +starlette==0.46.2 +scipy==1.15.2 +torchvision==0.22.0 +tornado==6.4.2 +cachetools==5.5.2 +platformdirs==4.3.7 +python-dotenv==1.1.0 +nvidia-cusolver-cu12==11.7.1.2 +referencing==0.36.2 +itsdangerous==2.2.0 +safetensors==0.5.3 +traitlets==5.14.3 +jupyterlab_server==2.27.3 +smolagents==1.16.1 +jupyterlab==4.4.2 +fqdn==1.5.1 +Werkzeug==3.1.3 +jupyter_client==8.6.3 +nvidia-nvtx-cu12==12.6.77 +grpcio==1.71.0 +sniffio==1.3.1 +aiosignal==1.3.2 +pillow==11.2.1 +triton==3.3.0 +greenlet==3.2.2 +orjson==3.10.18 +SQLAlchemy==2.0.41 +typing_extensions==4.13.2 +cloudpickle==3.1.1 +fastjsonschema==2.21.1 +transformers==4.52.4 +uvicorn==0.34.2 +typing-inspect==0.9.0 +pyzmq==26.4.0 +parso==0.8.4 +langchain==0.3.25 +yarl==1.20.0 +MarkupSafe==3.0.2 +mlflow-skinny==2.22.0 +tensorboard-data-server==0.7.2 +contourpy==1.3.2 +jupyter-events==0.12.0 +Mako==1.3.10 +wcwidth==0.2.13 +pyparsing==3.2.3 +Jinja2==3.1.6 +blinker==1.9.0 +json5==0.12.0 +opencv-python-headless==4.11.0.86 +ipywidgets==8.1.7 +packaging==25.0 +nbformat==5.10.4 +pytz==2025.2 +async-lru==2.0.5 +absl-py==2.2.2 +beautifulsoup4==4.13.4 +ipython==9.3.0 +executing==2.2.0 +cycler==0.12.1 +pydantic==2.11.5 +torchinfo==1.8.0 +math-verify==0.7.0 +regex==2024.11.6 +annotated-types==0.7.0 +smmap==5.0.2 diff --git a/wandb/offline-run-20250613_021623-6ot3basl/files/wandb-metadata.json b/wandb/offline-run-20250613_021623-6ot3basl/files/wandb-metadata.json new file mode 100644 index 0000000..f169ea6 --- /dev/null +++ b/wandb/offline-run-20250613_021623-6ot3basl/files/wandb-metadata.json @@ -0,0 +1,42 @@ +{ + "os": "Linux-6.11.0-26-generic-x86_64-with-glibc2.39", + "python": "CPython 3.12.3", + "startedAt": "2025-06-12T20:46:23.804504Z", + "args": [ + "--model_id", + "google/gemma-3-4b-pt", + "--dataset_id", + "ariG23498/license-detection-paligemma", + "--batch_size", + "8", + "--lr", + "2e-5", + "--checkpoint_id", + "oreonmayo/gemma-3-4b-pt-object-detection-aug" + ], + "program": "/home/manodeep/projects/NN/neural_networks_repo/finetuning/gemma-3_lora/gemma3-object-detection/train.py", + "codePath": "train.py", + "git": { + "remote": "https://github.com/Manodeepray/gemma3-object-detection", + "commit": "4d0d9cbaaaf2d548d7c42c7e4ec546c0d85e7143" + }, + "root": "/home/manodeep/projects/NN/neural_networks_repo/finetuning/gemma-3_lora/gemma3-object-detection", + "host": "mayo", + "executable": "/home/manodeep/projects/NN/.venv/bin/python", + "codePathLocal": "train.py", + "cpu_count": 4, + "cpu_count_logical": 8, + "disk": { + "/": { + "total": "208557645824", + "used": "87172231168" + } + }, + "memory": { + "total": "16297222144" + }, + "cpu": { + "count": 4, + "countLogical": 8 + } +} \ No newline at end of file diff --git a/wandb/offline-run-20250613_021623-6ot3basl/logs/debug-core.log b/wandb/offline-run-20250613_021623-6ot3basl/logs/debug-core.log new file mode 120000 index 0000000..775ae81 --- /dev/null +++ b/wandb/offline-run-20250613_021623-6ot3basl/logs/debug-core.log @@ -0,0 +1 @@ +/home/manodeep/.cache/wandb/logs/core-debug-20250613_021623.log \ No newline at end of file diff --git a/wandb/offline-run-20250613_021623-6ot3basl/logs/debug-internal.log b/wandb/offline-run-20250613_021623-6ot3basl/logs/debug-internal.log new file mode 100644 index 0000000..47d36af --- /dev/null +++ b/wandb/offline-run-20250613_021623-6ot3basl/logs/debug-internal.log @@ -0,0 +1,8 @@ +{"time":"2025-06-13T02:16:23.812623793+05:30","level":"INFO","msg":"stream: starting","core version":"0.20.1","symlink path":"/home/manodeep/projects/NN/neural_networks_repo/finetuning/gemma-3_lora/gemma3-object-detection/wandb/offline-run-20250613_021623-6ot3basl/logs/debug-core.log"} +{"time":"2025-06-13T02:16:23.92172777+05:30","level":"WARN","msg":"GraphQL client is nil, skipping feature loading"} +{"time":"2025-06-13T02:16:23.922117768+05:30","level":"INFO","msg":"stream: created new stream","id":"6ot3basl"} +{"time":"2025-06-13T02:16:23.922151309+05:30","level":"INFO","msg":"stream: started","id":"6ot3basl"} +{"time":"2025-06-13T02:16:23.922266365+05:30","level":"INFO","msg":"writer: Do: started","stream_id":"6ot3basl"} +{"time":"2025-06-13T02:16:23.922377998+05:30","level":"INFO","msg":"handler: started","stream_id":"6ot3basl"} +{"time":"2025-06-13T02:16:23.922420529+05:30","level":"INFO","msg":"sender: started","stream_id":"6ot3basl"} +{"time":"2025-06-13T02:16:23.928666516+05:30","level":"INFO","msg":"Starting system monitor"} diff --git a/wandb/offline-run-20250613_021623-6ot3basl/logs/debug.log b/wandb/offline-run-20250613_021623-6ot3basl/logs/debug.log new file mode 100644 index 0000000..180d9a9 --- /dev/null +++ b/wandb/offline-run-20250613_021623-6ot3basl/logs/debug.log @@ -0,0 +1,21 @@ +2025-06-13 02:16:23,564 INFO MainThread:53739 [wandb_setup.py:_flush():81] Current SDK version is 0.20.1 +2025-06-13 02:16:23,565 INFO MainThread:53739 [wandb_setup.py:_flush():81] Configure stats pid to 53739 +2025-06-13 02:16:23,565 INFO MainThread:53739 [wandb_setup.py:_flush():81] Loading settings from /home/manodeep/.config/wandb/settings +2025-06-13 02:16:23,565 INFO MainThread:53739 [wandb_setup.py:_flush():81] Loading settings from /home/manodeep/projects/NN/neural_networks_repo/finetuning/gemma-3_lora/gemma3-object-detection/wandb/settings +2025-06-13 02:16:23,565 INFO MainThread:53739 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-06-13 02:16:23,565 INFO MainThread:53739 [wandb_init.py:setup_run_log_directory():703] Logging user logs to /home/manodeep/projects/NN/neural_networks_repo/finetuning/gemma-3_lora/gemma3-object-detection/wandb/offline-run-20250613_021623-6ot3basl/logs/debug.log +2025-06-13 02:16:23,565 INFO MainThread:53739 [wandb_init.py:setup_run_log_directory():704] Logging internal logs to /home/manodeep/projects/NN/neural_networks_repo/finetuning/gemma-3_lora/gemma3-object-detection/wandb/offline-run-20250613_021623-6ot3basl/logs/debug-internal.log +2025-06-13 02:16:23,565 INFO MainThread:53739 [wandb_init.py:init():831] calling init triggers +2025-06-13 02:16:23,565 INFO MainThread:53739 [wandb_init.py:init():836] wandb.init called with sweep_config: {} +config: {'dataset_id': 'ariG23498/license-detection-paligemma', 'model_id': 'google/gemma-3-4b-pt', 'checkpoint_id': 'sergiopaniego/gemma-3-4b-pt-object-detection-aug', 'device': 'cpu', 'dtype': torch.bfloat16, 'batch_size': 8, 'learning_rate': 2e-05, 'project_name': 'train run', '_wandb': {}} +2025-06-13 02:16:23,565 INFO MainThread:53739 [wandb_init.py:init():872] starting backend +2025-06-13 02:16:23,788 INFO MainThread:53739 [wandb_init.py:init():875] sending inform_init request +2025-06-13 02:16:23,804 INFO MainThread:53739 [wandb_init.py:init():883] backend started and connected +2025-06-13 02:16:23,807 INFO MainThread:53739 [wandb_init.py:init():956] updated telemetry +2025-06-13 02:16:23,812 INFO MainThread:53739 [wandb_init.py:init():980] communicating run to backend with 90.0 second timeout +2025-06-13 02:16:23,924 INFO MainThread:53739 [wandb_init.py:init():1032] starting run threads in backend +2025-06-13 02:16:24,081 INFO MainThread:53739 [wandb_run.py:_console_start():2453] atexit reg +2025-06-13 02:16:24,081 INFO MainThread:53739 [wandb_run.py:_redirect():2301] redirect: wrap_raw +2025-06-13 02:16:24,081 INFO MainThread:53739 [wandb_run.py:_redirect():2370] Wrapping output streams. +2025-06-13 02:16:24,082 INFO MainThread:53739 [wandb_run.py:_redirect():2393] Redirects installed. +2025-06-13 02:16:24,085 INFO MainThread:53739 [wandb_init.py:init():1078] run started, returning control to user process diff --git a/wandb/offline-run-20250613_021623-6ot3basl/run-6ot3basl.wandb b/wandb/offline-run-20250613_021623-6ot3basl/run-6ot3basl.wandb new file mode 100644 index 0000000..e69de29 diff --git a/wandb/offline-run-20250613_022053-ok4hunfv/files/output.log b/wandb/offline-run-20250613_022053-ok4hunfv/files/output.log new file mode 100644 index 0000000..e69de29 diff --git a/wandb/offline-run-20250613_022053-ok4hunfv/files/requirements.txt b/wandb/offline-run-20250613_022053-ok4hunfv/files/requirements.txt new file mode 100644 index 0000000..2b4eca9 --- /dev/null +++ b/wandb/offline-run-20250613_022053-ok4hunfv/files/requirements.txt @@ -0,0 +1,253 @@ +Deprecated==1.2.18 +zipp==3.22.0 +langgraph-sdk==0.1.70 +psutil==7.0.0 +arrow==1.3.0 +aiohappyeyeballs==2.6.1 +charset-normalizer==3.4.2 +idna==3.10 +ptyprocess==0.7.0 +jupyter-lsp==2.2.5 +certifi==2025.4.26 +GitPython==3.1.44 +argparse==1.4.0 +webcolors==24.11.1 +bitsandbytes==0.46.0 +pandocfilters==1.5.1 +opentelemetry-sdk==1.33.1 +Flask==3.1.1 +ipython_pygments_lexers==1.1.1 +overrides==7.7.0 +notebook==7.4.2 +sympy==1.14.0 +cffi==1.17.1 +Send2Trash==1.8.3 +seaborn==0.13.2 +jedi==0.19.2 +anyio==4.9.0 +uri-template==1.3.0 +kiwisolver==1.4.8 +importlib_metadata==8.6.1 +pyppeteer==2.0.0 +stack-data==0.6.3 +comm==0.2.2 +tensorboard==2.19.0 +mistune==3.1.3 +isoduration==20.11.0 +nvidia-cusparse-cu12==12.5.4.2 +argon2-cffi-bindings==21.2.0 +accelerate==1.7.0 +wrapt==1.17.2 +stringzilla==3.12.5 +httpcore==1.0.9 +distro==1.9.0 +rpds-py==0.24.0 +python-json-logger==3.3.0 +pyasn1_modules==0.4.2 +requests-toolbelt==1.0.0 +Markdown==3.8 +aiohttp==3.12.6 +rsa==4.9.1 +albucore==0.0.24 +trl==0.18.1 +graphene==3.4.3 +nvidia-cublas-cu12==12.6.4.1 +typing-inspection==0.4.1 +threadpoolctl==3.6.0 +protobuf==6.31.0 +zstandard==0.23.0 +pandas==2.2.3 +langchain-groq==0.3.2 +nbconvert==7.16.6 +multidict==6.4.4 +babel==2.17.0 +opentelemetry-api==1.33.1 +torchaudio==2.7.0 +tzdata==2025.2 +dataclasses-json==0.6.7 +jsonschema==4.23.0 +matplotlib-inline==0.1.7 +networkx==3.5 +setuptools==80.9.0 +mpmath==1.3.0 +alembic==1.16.1 +prometheus_client==0.21.1 +argon2-cffi==23.1.0 +nvidia-cusparselt-cu12==0.6.3 +frozenlist==1.6.0 +pycparser==2.22 +jupyterlab_widgets==3.0.15 +multiprocess==0.70.16 +nvidia-curand-cu12==10.3.7.77 +langchain-community==0.3.24 +xxhash==3.5.0 +websocket-client==1.8.0 +soupsieve==2.7 +gitdb==4.0.12 +notebook_shim==0.2.4 +fonttools==4.57.0 +ipykernel==6.29.5 +graphql-core==3.2.6 +langchain-text-splitters==0.3.8 +langchain-core==0.3.61 +scikit-learn==1.6.1 +ormsgpack==1.9.1 +jupyter_core==5.7.2 +jupyter_server_terminals==0.5.3 +hf-xet==1.1.2 +jsonschema-specifications==2025.4.1 +opentelemetry-semantic-conventions==0.54b1 +nvidia-cuda-runtime-cu12==12.6.77 +tokenizers==0.21.1 +huggingface-hub==0.32.3 +torch==2.7.0 +nbclient==0.10.2 +websockets==10.4 +google-auth==2.40.2 +databricks-sdk==0.54.0 +fsspec==2025.3.0 +setproctitle==1.3.6 +PyYAML==6.0.2 +nvidia-cudnn-cu12==9.5.1.17 +marshmallow==3.26.1 +requests==2.32.3 +pyarrow==20.0.0 +langgraph==0.4.5 +pydantic_core==2.33.2 +langgraph-prebuilt==0.2.1 +prompt_toolkit==3.0.51 +nvidia-nvjitlink-cu12==12.6.85 +appdirs==1.4.4 +tinycss2==1.4.0 +asttokens==3.0.0 +wandb==0.20.1 +antlr4-python3-runtime==4.13.2 +tqdm==4.67.1 +bash_kernel==0.10.0 +nvidia-nccl-cu12==2.26.2 +numpy==2.2.6 +fastapi==0.115.12 +gunicorn==23.0.0 +attrs==25.3.0 +latex2sympy2_extended==1.10.1 +urllib3==2.4.0 +jsonpointer==3.0.0 +matplotlib==3.10.1 +pyee==11.1.1 +nvidia-cuda-cupti-cu12==12.6.80 +markdown-it-py==3.0.0 +httpx==0.28.1 +nest-asyncio==1.6.0 +pyasn1==0.6.1 +datasets==3.6.0 +decorator==5.2.1 +mdurl==0.1.2 +dill==0.3.8 +nvidia-cufft-cu12==11.3.0.4 +jupyter_server==2.15.0 +click==8.2.1 +jsonpatch==1.33 +webencodings==0.5.1 +defusedxml==0.7.1 +simsimd==6.4.9 +terminado==0.18.1 +tiktoken==0.9.0 +graphql-relay==3.2.0 +pexpect==4.9.0 +joblib==1.4.2 +rfc3339-validator==0.1.4 +rich==14.0.0 +tenacity==9.1.2 +widgetsnbextension==4.0.14 +h11==0.16.0 +debugpy==1.8.14 +nvidia-cufile-cu12==1.11.1.6 +filetype==1.2.0 +nvidia-cuda-nvrtc-cu12==12.6.77 +types-python-dateutil==2.9.0.20241206 +albumentations==2.0.8 +peft==0.15.2 +python-dateutil==2.9.0.post0 +rfc3986-validator==0.1.1 +bleach==6.2.0 +Pygments==2.19.1 +sentry-sdk==2.30.0 +propcache==0.3.1 +langsmith==0.3.42 +jupyterlab_pygments==0.3.0 +docker==7.1.0 +mypy_extensions==1.1.0 +groq==0.25.0 +mlflow==2.22.0 +pydantic-settings==2.9.1 +langgraph-checkpoint==2.0.26 +pure_eval==0.2.3 +sqlparse==0.5.3 +httpx-sse==0.4.0 +six==1.17.0 +filelock==3.18.0 +starlette==0.46.2 +scipy==1.15.2 +torchvision==0.22.0 +tornado==6.4.2 +cachetools==5.5.2 +platformdirs==4.3.7 +python-dotenv==1.1.0 +nvidia-cusolver-cu12==11.7.1.2 +referencing==0.36.2 +itsdangerous==2.2.0 +safetensors==0.5.3 +traitlets==5.14.3 +jupyterlab_server==2.27.3 +smolagents==1.16.1 +jupyterlab==4.4.2 +fqdn==1.5.1 +Werkzeug==3.1.3 +jupyter_client==8.6.3 +nvidia-nvtx-cu12==12.6.77 +grpcio==1.71.0 +sniffio==1.3.1 +aiosignal==1.3.2 +pillow==11.2.1 +triton==3.3.0 +greenlet==3.2.2 +orjson==3.10.18 +SQLAlchemy==2.0.41 +typing_extensions==4.13.2 +cloudpickle==3.1.1 +fastjsonschema==2.21.1 +transformers==4.52.4 +uvicorn==0.34.2 +typing-inspect==0.9.0 +pyzmq==26.4.0 +parso==0.8.4 +langchain==0.3.25 +yarl==1.20.0 +MarkupSafe==3.0.2 +mlflow-skinny==2.22.0 +tensorboard-data-server==0.7.2 +contourpy==1.3.2 +jupyter-events==0.12.0 +Mako==1.3.10 +wcwidth==0.2.13 +pyparsing==3.2.3 +Jinja2==3.1.6 +blinker==1.9.0 +json5==0.12.0 +opencv-python-headless==4.11.0.86 +ipywidgets==8.1.7 +packaging==25.0 +nbformat==5.10.4 +pytz==2025.2 +async-lru==2.0.5 +absl-py==2.2.2 +beautifulsoup4==4.13.4 +ipython==9.3.0 +executing==2.2.0 +cycler==0.12.1 +pydantic==2.11.5 +torchinfo==1.8.0 +math-verify==0.7.0 +regex==2024.11.6 +annotated-types==0.7.0 +smmap==5.0.2 diff --git a/wandb/offline-run-20250613_022053-ok4hunfv/files/wandb-metadata.json b/wandb/offline-run-20250613_022053-ok4hunfv/files/wandb-metadata.json new file mode 100644 index 0000000..06cfe18 --- /dev/null +++ b/wandb/offline-run-20250613_022053-ok4hunfv/files/wandb-metadata.json @@ -0,0 +1,46 @@ +{ + "os": "Linux-6.11.0-26-generic-x86_64-with-glibc2.39", + "python": "CPython 3.12.3", + "startedAt": "2025-06-12T20:50:53.785044Z", + "args": [ + "--model_id", + "google/gemma-3-4b-pt", + "--dataset_id", + "ariG23498/license-detection-paligemma", + "--batch_size", + "8", + "--lr", + "2e-5", + "--checkpoint_id", + "oreonmayo/gemma-3-4b-pt-object-detection-aug", + "--peft_type", + "lora", + "--peft_config", + "peft_configs/lora_configs.yaml" + ], + "program": "/home/manodeep/projects/NN/neural_networks_repo/finetuning/gemma-3_lora/gemma3-object-detection/finetune.py", + "codePath": "finetune.py", + "git": { + "remote": "https://github.com/Manodeepray/gemma3-object-detection", + "commit": "4d0d9cbaaaf2d548d7c42c7e4ec546c0d85e7143" + }, + "root": "/home/manodeep/projects/NN/neural_networks_repo/finetuning/gemma-3_lora/gemma3-object-detection", + "host": "mayo", + "executable": "/home/manodeep/projects/NN/.venv/bin/python", + "codePathLocal": "finetune.py", + "cpu_count": 4, + "cpu_count_logical": 8, + "disk": { + "/": { + "total": "208557645824", + "used": "87174987776" + } + }, + "memory": { + "total": "16297222144" + }, + "cpu": { + "count": 4, + "countLogical": 8 + } +} \ No newline at end of file diff --git a/wandb/offline-run-20250613_022053-ok4hunfv/logs/debug-core.log b/wandb/offline-run-20250613_022053-ok4hunfv/logs/debug-core.log new file mode 120000 index 0000000..592efb5 --- /dev/null +++ b/wandb/offline-run-20250613_022053-ok4hunfv/logs/debug-core.log @@ -0,0 +1 @@ +/home/manodeep/.cache/wandb/logs/core-debug-20250613_022053.log \ No newline at end of file diff --git a/wandb/offline-run-20250613_022053-ok4hunfv/logs/debug-internal.log b/wandb/offline-run-20250613_022053-ok4hunfv/logs/debug-internal.log new file mode 100644 index 0000000..bf7a12d --- /dev/null +++ b/wandb/offline-run-20250613_022053-ok4hunfv/logs/debug-internal.log @@ -0,0 +1,8 @@ +{"time":"2025-06-13T02:20:53.797779656+05:30","level":"INFO","msg":"stream: starting","core version":"0.20.1","symlink path":"/home/manodeep/projects/NN/neural_networks_repo/finetuning/gemma-3_lora/gemma3-object-detection/wandb/offline-run-20250613_022053-ok4hunfv/logs/debug-core.log"} +{"time":"2025-06-13T02:20:53.903803604+05:30","level":"WARN","msg":"GraphQL client is nil, skipping feature loading"} +{"time":"2025-06-13T02:20:53.904210076+05:30","level":"INFO","msg":"stream: created new stream","id":"ok4hunfv"} +{"time":"2025-06-13T02:20:53.904243526+05:30","level":"INFO","msg":"stream: started","id":"ok4hunfv"} +{"time":"2025-06-13T02:20:53.904330248+05:30","level":"INFO","msg":"sender: started","stream_id":"ok4hunfv"} +{"time":"2025-06-13T02:20:53.904303338+05:30","level":"INFO","msg":"writer: Do: started","stream_id":"ok4hunfv"} +{"time":"2025-06-13T02:20:53.904357482+05:30","level":"INFO","msg":"handler: started","stream_id":"ok4hunfv"} +{"time":"2025-06-13T02:20:53.911351481+05:30","level":"INFO","msg":"Starting system monitor"} diff --git a/wandb/offline-run-20250613_022053-ok4hunfv/logs/debug.log b/wandb/offline-run-20250613_022053-ok4hunfv/logs/debug.log new file mode 100644 index 0000000..ccf00f8 --- /dev/null +++ b/wandb/offline-run-20250613_022053-ok4hunfv/logs/debug.log @@ -0,0 +1,21 @@ +2025-06-13 02:20:53,564 INFO MainThread:54298 [wandb_setup.py:_flush():81] Current SDK version is 0.20.1 +2025-06-13 02:20:53,564 INFO MainThread:54298 [wandb_setup.py:_flush():81] Configure stats pid to 54298 +2025-06-13 02:20:53,564 INFO MainThread:54298 [wandb_setup.py:_flush():81] Loading settings from /home/manodeep/.config/wandb/settings +2025-06-13 02:20:53,564 INFO MainThread:54298 [wandb_setup.py:_flush():81] Loading settings from /home/manodeep/projects/NN/neural_networks_repo/finetuning/gemma-3_lora/gemma3-object-detection/wandb/settings +2025-06-13 02:20:53,564 INFO MainThread:54298 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-06-13 02:20:53,564 INFO MainThread:54298 [wandb_init.py:setup_run_log_directory():703] Logging user logs to /home/manodeep/projects/NN/neural_networks_repo/finetuning/gemma-3_lora/gemma3-object-detection/wandb/offline-run-20250613_022053-ok4hunfv/logs/debug.log +2025-06-13 02:20:53,564 INFO MainThread:54298 [wandb_init.py:setup_run_log_directory():704] Logging internal logs to /home/manodeep/projects/NN/neural_networks_repo/finetuning/gemma-3_lora/gemma3-object-detection/wandb/offline-run-20250613_022053-ok4hunfv/logs/debug-internal.log +2025-06-13 02:20:53,564 INFO MainThread:54298 [wandb_init.py:init():831] calling init triggers +2025-06-13 02:20:53,564 INFO MainThread:54298 [wandb_init.py:init():836] wandb.init called with sweep_config: {} +config: {'dataset_id': 'ariG23498/license-detection-paligemma', 'model_id': 'google/gemma-3-4b-pt', 'checkpoint_id': 'sergiopaniego/gemma-3-4b-pt-object-detection-aug', 'device': 'cpu', 'dtype': torch.bfloat16, 'batch_size': 8, 'learning_rate': 2e-05, 'project_name': 'gemma_finetune', '_wandb': {}} +2025-06-13 02:20:53,564 INFO MainThread:54298 [wandb_init.py:init():872] starting backend +2025-06-13 02:20:53,775 INFO MainThread:54298 [wandb_init.py:init():875] sending inform_init request +2025-06-13 02:20:53,784 INFO MainThread:54298 [wandb_init.py:init():883] backend started and connected +2025-06-13 02:20:53,786 INFO MainThread:54298 [wandb_init.py:init():956] updated telemetry +2025-06-13 02:20:53,795 INFO MainThread:54298 [wandb_init.py:init():980] communicating run to backend with 90.0 second timeout +2025-06-13 02:20:53,907 INFO MainThread:54298 [wandb_init.py:init():1032] starting run threads in backend +2025-06-13 02:20:54,061 INFO MainThread:54298 [wandb_run.py:_console_start():2453] atexit reg +2025-06-13 02:20:54,061 INFO MainThread:54298 [wandb_run.py:_redirect():2301] redirect: wrap_raw +2025-06-13 02:20:54,061 INFO MainThread:54298 [wandb_run.py:_redirect():2370] Wrapping output streams. +2025-06-13 02:20:54,061 INFO MainThread:54298 [wandb_run.py:_redirect():2393] Redirects installed. +2025-06-13 02:20:54,062 INFO MainThread:54298 [wandb_init.py:init():1078] run started, returning control to user process diff --git a/wandb/offline-run-20250613_022053-ok4hunfv/run-ok4hunfv.wandb b/wandb/offline-run-20250613_022053-ok4hunfv/run-ok4hunfv.wandb new file mode 100644 index 0000000..e69de29 From 7896d84d8f120d7f2167ae23c04a619c6125f887 Mon Sep 17 00:00:00 2001 From: Manodeepray Date: Fri, 13 Jun 2025 02:22:33 +0530 Subject: [PATCH 3/5] Fix: Resolve issue #12 --- .gitignore | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/.gitignore b/.gitignore index 95c7f21..762728c 100644 --- a/.gitignore +++ b/.gitignore @@ -1,3 +1,4 @@ .ruff_cache .venv -__pycache__ \ No newline at end of file +__pycache__ +wandb \ No newline at end of file From 6853a1a5fcd99d93d78e8702fb2faee099b21cf7 Mon Sep 17 00:00:00 2001 From: Manodeepray Date: Fri, 13 Jun 2025 02:25:00 +0530 Subject: [PATCH 4/5] Fix: Resolve issue #12 --- finetune.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/finetune.py b/finetune.py index 87fb680..1555a9e 100644 --- a/finetune.py +++ b/finetune.py @@ -183,7 +183,7 @@ def get_peft_config(peft_type: str, config_dict: dict) -> LoraConfig: wandb.init( project=cfg.project_name, - name=cfg.run_name if hasattr(cfg, "run_name") else None, + name=cfg.run_name if hasattr(cfg, "run_name") else f"{args.peft_type} run", config=vars(cfg), ) From 0626aefe3eb4a1a6ad8eaef98d3b230597a0be1c Mon Sep 17 00:00:00 2001 From: Manodeepray Date: Fri, 13 Jun 2025 02:56:17 +0530 Subject: [PATCH 5/5] Fix: Resolve issue #12 --- wandb/debug-internal.log | 1 - wandb/debug.log | 1 - wandb/latest-run | 1 - .../files/output.log | 0 .../files/requirements.txt | 253 ------------------ .../files/wandb-metadata.json | 42 --- .../logs/debug-core.log | 1 - .../logs/debug-internal.log | 8 - .../logs/debug.log | 21 -- .../run-6ot3basl.wandb | 0 .../files/output.log | 0 .../files/requirements.txt | 253 ------------------ .../files/wandb-metadata.json | 46 ---- .../logs/debug-core.log | 1 - .../logs/debug-internal.log | 8 - .../logs/debug.log | 21 -- .../run-ok4hunfv.wandb | 0 17 files changed, 657 deletions(-) delete mode 120000 wandb/debug-internal.log delete mode 120000 wandb/debug.log delete mode 120000 wandb/latest-run delete mode 100644 wandb/offline-run-20250613_021623-6ot3basl/files/output.log delete mode 100644 wandb/offline-run-20250613_021623-6ot3basl/files/requirements.txt delete mode 100644 wandb/offline-run-20250613_021623-6ot3basl/files/wandb-metadata.json delete mode 120000 wandb/offline-run-20250613_021623-6ot3basl/logs/debug-core.log delete mode 100644 wandb/offline-run-20250613_021623-6ot3basl/logs/debug-internal.log delete mode 100644 wandb/offline-run-20250613_021623-6ot3basl/logs/debug.log delete mode 100644 wandb/offline-run-20250613_021623-6ot3basl/run-6ot3basl.wandb delete mode 100644 wandb/offline-run-20250613_022053-ok4hunfv/files/output.log delete mode 100644 wandb/offline-run-20250613_022053-ok4hunfv/files/requirements.txt delete mode 100644 wandb/offline-run-20250613_022053-ok4hunfv/files/wandb-metadata.json delete mode 120000 wandb/offline-run-20250613_022053-ok4hunfv/logs/debug-core.log delete mode 100644 wandb/offline-run-20250613_022053-ok4hunfv/logs/debug-internal.log delete mode 100644 wandb/offline-run-20250613_022053-ok4hunfv/logs/debug.log delete mode 100644 wandb/offline-run-20250613_022053-ok4hunfv/run-ok4hunfv.wandb diff --git a/wandb/debug-internal.log b/wandb/debug-internal.log deleted file mode 120000 index 6955c19..0000000 --- a/wandb/debug-internal.log +++ /dev/null @@ -1 +0,0 @@ -offline-run-20250613_022053-ok4hunfv/logs/debug-internal.log \ No newline at end of file diff --git a/wandb/debug.log b/wandb/debug.log deleted file mode 120000 index 6b93e96..0000000 --- a/wandb/debug.log +++ /dev/null @@ -1 +0,0 @@ -offline-run-20250613_022053-ok4hunfv/logs/debug.log \ No newline at end of file diff --git a/wandb/latest-run b/wandb/latest-run deleted file mode 120000 index 430abf7..0000000 --- a/wandb/latest-run +++ /dev/null @@ -1 +0,0 @@ -offline-run-20250613_022053-ok4hunfv \ No newline at end of file diff --git a/wandb/offline-run-20250613_021623-6ot3basl/files/output.log b/wandb/offline-run-20250613_021623-6ot3basl/files/output.log deleted file mode 100644 index e69de29..0000000 diff --git a/wandb/offline-run-20250613_021623-6ot3basl/files/requirements.txt b/wandb/offline-run-20250613_021623-6ot3basl/files/requirements.txt deleted file mode 100644 index 2b4eca9..0000000 --- a/wandb/offline-run-20250613_021623-6ot3basl/files/requirements.txt +++ /dev/null @@ -1,253 +0,0 @@ -Deprecated==1.2.18 -zipp==3.22.0 -langgraph-sdk==0.1.70 -psutil==7.0.0 -arrow==1.3.0 -aiohappyeyeballs==2.6.1 -charset-normalizer==3.4.2 -idna==3.10 -ptyprocess==0.7.0 -jupyter-lsp==2.2.5 -certifi==2025.4.26 -GitPython==3.1.44 -argparse==1.4.0 -webcolors==24.11.1 -bitsandbytes==0.46.0 -pandocfilters==1.5.1 -opentelemetry-sdk==1.33.1 -Flask==3.1.1 -ipython_pygments_lexers==1.1.1 -overrides==7.7.0 -notebook==7.4.2 -sympy==1.14.0 -cffi==1.17.1 -Send2Trash==1.8.3 -seaborn==0.13.2 -jedi==0.19.2 -anyio==4.9.0 -uri-template==1.3.0 -kiwisolver==1.4.8 -importlib_metadata==8.6.1 -pyppeteer==2.0.0 -stack-data==0.6.3 -comm==0.2.2 -tensorboard==2.19.0 -mistune==3.1.3 -isoduration==20.11.0 -nvidia-cusparse-cu12==12.5.4.2 -argon2-cffi-bindings==21.2.0 -accelerate==1.7.0 -wrapt==1.17.2 -stringzilla==3.12.5 -httpcore==1.0.9 -distro==1.9.0 -rpds-py==0.24.0 -python-json-logger==3.3.0 -pyasn1_modules==0.4.2 -requests-toolbelt==1.0.0 -Markdown==3.8 -aiohttp==3.12.6 -rsa==4.9.1 -albucore==0.0.24 -trl==0.18.1 -graphene==3.4.3 -nvidia-cublas-cu12==12.6.4.1 -typing-inspection==0.4.1 -threadpoolctl==3.6.0 -protobuf==6.31.0 -zstandard==0.23.0 -pandas==2.2.3 -langchain-groq==0.3.2 -nbconvert==7.16.6 -multidict==6.4.4 -babel==2.17.0 -opentelemetry-api==1.33.1 -torchaudio==2.7.0 -tzdata==2025.2 -dataclasses-json==0.6.7 -jsonschema==4.23.0 -matplotlib-inline==0.1.7 -networkx==3.5 -setuptools==80.9.0 -mpmath==1.3.0 -alembic==1.16.1 -prometheus_client==0.21.1 -argon2-cffi==23.1.0 -nvidia-cusparselt-cu12==0.6.3 -frozenlist==1.6.0 -pycparser==2.22 -jupyterlab_widgets==3.0.15 -multiprocess==0.70.16 -nvidia-curand-cu12==10.3.7.77 -langchain-community==0.3.24 -xxhash==3.5.0 -websocket-client==1.8.0 -soupsieve==2.7 -gitdb==4.0.12 -notebook_shim==0.2.4 -fonttools==4.57.0 -ipykernel==6.29.5 -graphql-core==3.2.6 -langchain-text-splitters==0.3.8 -langchain-core==0.3.61 -scikit-learn==1.6.1 -ormsgpack==1.9.1 -jupyter_core==5.7.2 -jupyter_server_terminals==0.5.3 -hf-xet==1.1.2 -jsonschema-specifications==2025.4.1 -opentelemetry-semantic-conventions==0.54b1 -nvidia-cuda-runtime-cu12==12.6.77 -tokenizers==0.21.1 -huggingface-hub==0.32.3 -torch==2.7.0 -nbclient==0.10.2 -websockets==10.4 -google-auth==2.40.2 -databricks-sdk==0.54.0 -fsspec==2025.3.0 -setproctitle==1.3.6 -PyYAML==6.0.2 -nvidia-cudnn-cu12==9.5.1.17 -marshmallow==3.26.1 -requests==2.32.3 -pyarrow==20.0.0 -langgraph==0.4.5 -pydantic_core==2.33.2 -langgraph-prebuilt==0.2.1 -prompt_toolkit==3.0.51 -nvidia-nvjitlink-cu12==12.6.85 -appdirs==1.4.4 -tinycss2==1.4.0 -asttokens==3.0.0 -wandb==0.20.1 -antlr4-python3-runtime==4.13.2 -tqdm==4.67.1 -bash_kernel==0.10.0 -nvidia-nccl-cu12==2.26.2 -numpy==2.2.6 -fastapi==0.115.12 -gunicorn==23.0.0 -attrs==25.3.0 -latex2sympy2_extended==1.10.1 -urllib3==2.4.0 -jsonpointer==3.0.0 -matplotlib==3.10.1 -pyee==11.1.1 -nvidia-cuda-cupti-cu12==12.6.80 -markdown-it-py==3.0.0 -httpx==0.28.1 -nest-asyncio==1.6.0 -pyasn1==0.6.1 -datasets==3.6.0 -decorator==5.2.1 -mdurl==0.1.2 -dill==0.3.8 -nvidia-cufft-cu12==11.3.0.4 -jupyter_server==2.15.0 -click==8.2.1 -jsonpatch==1.33 -webencodings==0.5.1 -defusedxml==0.7.1 -simsimd==6.4.9 -terminado==0.18.1 -tiktoken==0.9.0 -graphql-relay==3.2.0 -pexpect==4.9.0 -joblib==1.4.2 -rfc3339-validator==0.1.4 -rich==14.0.0 -tenacity==9.1.2 -widgetsnbextension==4.0.14 -h11==0.16.0 -debugpy==1.8.14 -nvidia-cufile-cu12==1.11.1.6 -filetype==1.2.0 -nvidia-cuda-nvrtc-cu12==12.6.77 -types-python-dateutil==2.9.0.20241206 -albumentations==2.0.8 -peft==0.15.2 -python-dateutil==2.9.0.post0 -rfc3986-validator==0.1.1 -bleach==6.2.0 -Pygments==2.19.1 -sentry-sdk==2.30.0 -propcache==0.3.1 -langsmith==0.3.42 -jupyterlab_pygments==0.3.0 -docker==7.1.0 -mypy_extensions==1.1.0 -groq==0.25.0 -mlflow==2.22.0 -pydantic-settings==2.9.1 -langgraph-checkpoint==2.0.26 -pure_eval==0.2.3 -sqlparse==0.5.3 -httpx-sse==0.4.0 -six==1.17.0 -filelock==3.18.0 -starlette==0.46.2 -scipy==1.15.2 -torchvision==0.22.0 -tornado==6.4.2 -cachetools==5.5.2 -platformdirs==4.3.7 -python-dotenv==1.1.0 -nvidia-cusolver-cu12==11.7.1.2 -referencing==0.36.2 -itsdangerous==2.2.0 -safetensors==0.5.3 -traitlets==5.14.3 -jupyterlab_server==2.27.3 -smolagents==1.16.1 -jupyterlab==4.4.2 -fqdn==1.5.1 -Werkzeug==3.1.3 -jupyter_client==8.6.3 -nvidia-nvtx-cu12==12.6.77 -grpcio==1.71.0 -sniffio==1.3.1 -aiosignal==1.3.2 -pillow==11.2.1 -triton==3.3.0 -greenlet==3.2.2 -orjson==3.10.18 -SQLAlchemy==2.0.41 -typing_extensions==4.13.2 -cloudpickle==3.1.1 -fastjsonschema==2.21.1 -transformers==4.52.4 -uvicorn==0.34.2 -typing-inspect==0.9.0 -pyzmq==26.4.0 -parso==0.8.4 -langchain==0.3.25 -yarl==1.20.0 -MarkupSafe==3.0.2 -mlflow-skinny==2.22.0 -tensorboard-data-server==0.7.2 -contourpy==1.3.2 -jupyter-events==0.12.0 -Mako==1.3.10 -wcwidth==0.2.13 -pyparsing==3.2.3 -Jinja2==3.1.6 -blinker==1.9.0 -json5==0.12.0 -opencv-python-headless==4.11.0.86 -ipywidgets==8.1.7 -packaging==25.0 -nbformat==5.10.4 -pytz==2025.2 -async-lru==2.0.5 -absl-py==2.2.2 -beautifulsoup4==4.13.4 -ipython==9.3.0 -executing==2.2.0 -cycler==0.12.1 -pydantic==2.11.5 -torchinfo==1.8.0 -math-verify==0.7.0 -regex==2024.11.6 -annotated-types==0.7.0 -smmap==5.0.2 diff --git a/wandb/offline-run-20250613_021623-6ot3basl/files/wandb-metadata.json b/wandb/offline-run-20250613_021623-6ot3basl/files/wandb-metadata.json deleted file mode 100644 index f169ea6..0000000 --- a/wandb/offline-run-20250613_021623-6ot3basl/files/wandb-metadata.json +++ /dev/null @@ -1,42 +0,0 @@ -{ - "os": "Linux-6.11.0-26-generic-x86_64-with-glibc2.39", - "python": "CPython 3.12.3", - "startedAt": "2025-06-12T20:46:23.804504Z", - "args": [ - "--model_id", - "google/gemma-3-4b-pt", - "--dataset_id", - "ariG23498/license-detection-paligemma", - "--batch_size", - "8", - "--lr", - "2e-5", - "--checkpoint_id", - "oreonmayo/gemma-3-4b-pt-object-detection-aug" - ], - "program": "/home/manodeep/projects/NN/neural_networks_repo/finetuning/gemma-3_lora/gemma3-object-detection/train.py", - "codePath": "train.py", - "git": { - "remote": "https://github.com/Manodeepray/gemma3-object-detection", - "commit": "4d0d9cbaaaf2d548d7c42c7e4ec546c0d85e7143" - }, - "root": "/home/manodeep/projects/NN/neural_networks_repo/finetuning/gemma-3_lora/gemma3-object-detection", - "host": "mayo", - "executable": "/home/manodeep/projects/NN/.venv/bin/python", - "codePathLocal": "train.py", - "cpu_count": 4, - "cpu_count_logical": 8, - "disk": { - "/": { - "total": "208557645824", - "used": "87172231168" - } - }, - "memory": { - "total": "16297222144" - }, - "cpu": { - "count": 4, - "countLogical": 8 - } -} \ No newline at end of file diff --git a/wandb/offline-run-20250613_021623-6ot3basl/logs/debug-core.log b/wandb/offline-run-20250613_021623-6ot3basl/logs/debug-core.log deleted file mode 120000 index 775ae81..0000000 --- a/wandb/offline-run-20250613_021623-6ot3basl/logs/debug-core.log +++ /dev/null @@ -1 +0,0 @@ -/home/manodeep/.cache/wandb/logs/core-debug-20250613_021623.log \ No newline at end of file diff --git a/wandb/offline-run-20250613_021623-6ot3basl/logs/debug-internal.log b/wandb/offline-run-20250613_021623-6ot3basl/logs/debug-internal.log deleted file mode 100644 index 47d36af..0000000 --- a/wandb/offline-run-20250613_021623-6ot3basl/logs/debug-internal.log +++ /dev/null @@ -1,8 +0,0 @@ -{"time":"2025-06-13T02:16:23.812623793+05:30","level":"INFO","msg":"stream: starting","core version":"0.20.1","symlink path":"/home/manodeep/projects/NN/neural_networks_repo/finetuning/gemma-3_lora/gemma3-object-detection/wandb/offline-run-20250613_021623-6ot3basl/logs/debug-core.log"} -{"time":"2025-06-13T02:16:23.92172777+05:30","level":"WARN","msg":"GraphQL client is nil, skipping feature loading"} -{"time":"2025-06-13T02:16:23.922117768+05:30","level":"INFO","msg":"stream: created new stream","id":"6ot3basl"} -{"time":"2025-06-13T02:16:23.922151309+05:30","level":"INFO","msg":"stream: started","id":"6ot3basl"} -{"time":"2025-06-13T02:16:23.922266365+05:30","level":"INFO","msg":"writer: Do: started","stream_id":"6ot3basl"} -{"time":"2025-06-13T02:16:23.922377998+05:30","level":"INFO","msg":"handler: started","stream_id":"6ot3basl"} -{"time":"2025-06-13T02:16:23.922420529+05:30","level":"INFO","msg":"sender: started","stream_id":"6ot3basl"} -{"time":"2025-06-13T02:16:23.928666516+05:30","level":"INFO","msg":"Starting system monitor"} diff --git a/wandb/offline-run-20250613_021623-6ot3basl/logs/debug.log b/wandb/offline-run-20250613_021623-6ot3basl/logs/debug.log deleted file mode 100644 index 180d9a9..0000000 --- a/wandb/offline-run-20250613_021623-6ot3basl/logs/debug.log +++ /dev/null @@ -1,21 +0,0 @@ -2025-06-13 02:16:23,564 INFO MainThread:53739 [wandb_setup.py:_flush():81] Current SDK version is 0.20.1 -2025-06-13 02:16:23,565 INFO MainThread:53739 [wandb_setup.py:_flush():81] Configure stats pid to 53739 -2025-06-13 02:16:23,565 INFO MainThread:53739 [wandb_setup.py:_flush():81] Loading settings from /home/manodeep/.config/wandb/settings -2025-06-13 02:16:23,565 INFO MainThread:53739 [wandb_setup.py:_flush():81] Loading settings from /home/manodeep/projects/NN/neural_networks_repo/finetuning/gemma-3_lora/gemma3-object-detection/wandb/settings -2025-06-13 02:16:23,565 INFO MainThread:53739 [wandb_setup.py:_flush():81] Loading settings from environment variables -2025-06-13 02:16:23,565 INFO MainThread:53739 [wandb_init.py:setup_run_log_directory():703] Logging user logs to /home/manodeep/projects/NN/neural_networks_repo/finetuning/gemma-3_lora/gemma3-object-detection/wandb/offline-run-20250613_021623-6ot3basl/logs/debug.log -2025-06-13 02:16:23,565 INFO MainThread:53739 [wandb_init.py:setup_run_log_directory():704] Logging internal logs to /home/manodeep/projects/NN/neural_networks_repo/finetuning/gemma-3_lora/gemma3-object-detection/wandb/offline-run-20250613_021623-6ot3basl/logs/debug-internal.log -2025-06-13 02:16:23,565 INFO MainThread:53739 [wandb_init.py:init():831] calling init triggers -2025-06-13 02:16:23,565 INFO MainThread:53739 [wandb_init.py:init():836] wandb.init called with sweep_config: {} -config: {'dataset_id': 'ariG23498/license-detection-paligemma', 'model_id': 'google/gemma-3-4b-pt', 'checkpoint_id': 'sergiopaniego/gemma-3-4b-pt-object-detection-aug', 'device': 'cpu', 'dtype': torch.bfloat16, 'batch_size': 8, 'learning_rate': 2e-05, 'project_name': 'train run', '_wandb': {}} -2025-06-13 02:16:23,565 INFO MainThread:53739 [wandb_init.py:init():872] starting backend -2025-06-13 02:16:23,788 INFO MainThread:53739 [wandb_init.py:init():875] sending inform_init request -2025-06-13 02:16:23,804 INFO MainThread:53739 [wandb_init.py:init():883] backend started and connected -2025-06-13 02:16:23,807 INFO MainThread:53739 [wandb_init.py:init():956] updated telemetry -2025-06-13 02:16:23,812 INFO MainThread:53739 [wandb_init.py:init():980] communicating run to backend with 90.0 second timeout -2025-06-13 02:16:23,924 INFO MainThread:53739 [wandb_init.py:init():1032] starting run threads in backend -2025-06-13 02:16:24,081 INFO MainThread:53739 [wandb_run.py:_console_start():2453] atexit reg -2025-06-13 02:16:24,081 INFO MainThread:53739 [wandb_run.py:_redirect():2301] redirect: wrap_raw -2025-06-13 02:16:24,081 INFO MainThread:53739 [wandb_run.py:_redirect():2370] Wrapping output streams. -2025-06-13 02:16:24,082 INFO MainThread:53739 [wandb_run.py:_redirect():2393] Redirects installed. -2025-06-13 02:16:24,085 INFO MainThread:53739 [wandb_init.py:init():1078] run started, returning control to user process diff --git a/wandb/offline-run-20250613_021623-6ot3basl/run-6ot3basl.wandb b/wandb/offline-run-20250613_021623-6ot3basl/run-6ot3basl.wandb deleted file mode 100644 index e69de29..0000000 diff --git a/wandb/offline-run-20250613_022053-ok4hunfv/files/output.log b/wandb/offline-run-20250613_022053-ok4hunfv/files/output.log deleted file mode 100644 index e69de29..0000000 diff --git a/wandb/offline-run-20250613_022053-ok4hunfv/files/requirements.txt b/wandb/offline-run-20250613_022053-ok4hunfv/files/requirements.txt deleted file mode 100644 index 2b4eca9..0000000 --- a/wandb/offline-run-20250613_022053-ok4hunfv/files/requirements.txt +++ /dev/null @@ -1,253 +0,0 @@ -Deprecated==1.2.18 -zipp==3.22.0 -langgraph-sdk==0.1.70 -psutil==7.0.0 -arrow==1.3.0 -aiohappyeyeballs==2.6.1 -charset-normalizer==3.4.2 -idna==3.10 -ptyprocess==0.7.0 -jupyter-lsp==2.2.5 -certifi==2025.4.26 -GitPython==3.1.44 -argparse==1.4.0 -webcolors==24.11.1 -bitsandbytes==0.46.0 -pandocfilters==1.5.1 -opentelemetry-sdk==1.33.1 -Flask==3.1.1 -ipython_pygments_lexers==1.1.1 -overrides==7.7.0 -notebook==7.4.2 -sympy==1.14.0 -cffi==1.17.1 -Send2Trash==1.8.3 -seaborn==0.13.2 -jedi==0.19.2 -anyio==4.9.0 -uri-template==1.3.0 -kiwisolver==1.4.8 -importlib_metadata==8.6.1 -pyppeteer==2.0.0 -stack-data==0.6.3 -comm==0.2.2 -tensorboard==2.19.0 -mistune==3.1.3 -isoduration==20.11.0 -nvidia-cusparse-cu12==12.5.4.2 -argon2-cffi-bindings==21.2.0 -accelerate==1.7.0 -wrapt==1.17.2 -stringzilla==3.12.5 -httpcore==1.0.9 -distro==1.9.0 -rpds-py==0.24.0 -python-json-logger==3.3.0 -pyasn1_modules==0.4.2 -requests-toolbelt==1.0.0 -Markdown==3.8 -aiohttp==3.12.6 -rsa==4.9.1 -albucore==0.0.24 -trl==0.18.1 -graphene==3.4.3 -nvidia-cublas-cu12==12.6.4.1 -typing-inspection==0.4.1 -threadpoolctl==3.6.0 -protobuf==6.31.0 -zstandard==0.23.0 -pandas==2.2.3 -langchain-groq==0.3.2 -nbconvert==7.16.6 -multidict==6.4.4 -babel==2.17.0 -opentelemetry-api==1.33.1 -torchaudio==2.7.0 -tzdata==2025.2 -dataclasses-json==0.6.7 -jsonschema==4.23.0 -matplotlib-inline==0.1.7 -networkx==3.5 -setuptools==80.9.0 -mpmath==1.3.0 -alembic==1.16.1 -prometheus_client==0.21.1 -argon2-cffi==23.1.0 -nvidia-cusparselt-cu12==0.6.3 -frozenlist==1.6.0 -pycparser==2.22 -jupyterlab_widgets==3.0.15 -multiprocess==0.70.16 -nvidia-curand-cu12==10.3.7.77 -langchain-community==0.3.24 -xxhash==3.5.0 -websocket-client==1.8.0 -soupsieve==2.7 -gitdb==4.0.12 -notebook_shim==0.2.4 -fonttools==4.57.0 -ipykernel==6.29.5 -graphql-core==3.2.6 -langchain-text-splitters==0.3.8 -langchain-core==0.3.61 -scikit-learn==1.6.1 -ormsgpack==1.9.1 -jupyter_core==5.7.2 -jupyter_server_terminals==0.5.3 -hf-xet==1.1.2 -jsonschema-specifications==2025.4.1 -opentelemetry-semantic-conventions==0.54b1 -nvidia-cuda-runtime-cu12==12.6.77 -tokenizers==0.21.1 -huggingface-hub==0.32.3 -torch==2.7.0 -nbclient==0.10.2 -websockets==10.4 -google-auth==2.40.2 -databricks-sdk==0.54.0 -fsspec==2025.3.0 -setproctitle==1.3.6 -PyYAML==6.0.2 -nvidia-cudnn-cu12==9.5.1.17 -marshmallow==3.26.1 -requests==2.32.3 -pyarrow==20.0.0 -langgraph==0.4.5 -pydantic_core==2.33.2 -langgraph-prebuilt==0.2.1 -prompt_toolkit==3.0.51 -nvidia-nvjitlink-cu12==12.6.85 -appdirs==1.4.4 -tinycss2==1.4.0 -asttokens==3.0.0 -wandb==0.20.1 -antlr4-python3-runtime==4.13.2 -tqdm==4.67.1 -bash_kernel==0.10.0 -nvidia-nccl-cu12==2.26.2 -numpy==2.2.6 -fastapi==0.115.12 -gunicorn==23.0.0 -attrs==25.3.0 -latex2sympy2_extended==1.10.1 -urllib3==2.4.0 -jsonpointer==3.0.0 -matplotlib==3.10.1 -pyee==11.1.1 -nvidia-cuda-cupti-cu12==12.6.80 -markdown-it-py==3.0.0 -httpx==0.28.1 -nest-asyncio==1.6.0 -pyasn1==0.6.1 -datasets==3.6.0 -decorator==5.2.1 -mdurl==0.1.2 -dill==0.3.8 -nvidia-cufft-cu12==11.3.0.4 -jupyter_server==2.15.0 -click==8.2.1 -jsonpatch==1.33 -webencodings==0.5.1 -defusedxml==0.7.1 -simsimd==6.4.9 -terminado==0.18.1 -tiktoken==0.9.0 -graphql-relay==3.2.0 -pexpect==4.9.0 -joblib==1.4.2 -rfc3339-validator==0.1.4 -rich==14.0.0 -tenacity==9.1.2 -widgetsnbextension==4.0.14 -h11==0.16.0 -debugpy==1.8.14 -nvidia-cufile-cu12==1.11.1.6 -filetype==1.2.0 -nvidia-cuda-nvrtc-cu12==12.6.77 -types-python-dateutil==2.9.0.20241206 -albumentations==2.0.8 -peft==0.15.2 -python-dateutil==2.9.0.post0 -rfc3986-validator==0.1.1 -bleach==6.2.0 -Pygments==2.19.1 -sentry-sdk==2.30.0 -propcache==0.3.1 -langsmith==0.3.42 -jupyterlab_pygments==0.3.0 -docker==7.1.0 -mypy_extensions==1.1.0 -groq==0.25.0 -mlflow==2.22.0 -pydantic-settings==2.9.1 -langgraph-checkpoint==2.0.26 -pure_eval==0.2.3 -sqlparse==0.5.3 -httpx-sse==0.4.0 -six==1.17.0 -filelock==3.18.0 -starlette==0.46.2 -scipy==1.15.2 -torchvision==0.22.0 -tornado==6.4.2 -cachetools==5.5.2 -platformdirs==4.3.7 -python-dotenv==1.1.0 -nvidia-cusolver-cu12==11.7.1.2 -referencing==0.36.2 -itsdangerous==2.2.0 -safetensors==0.5.3 -traitlets==5.14.3 -jupyterlab_server==2.27.3 -smolagents==1.16.1 -jupyterlab==4.4.2 -fqdn==1.5.1 -Werkzeug==3.1.3 -jupyter_client==8.6.3 -nvidia-nvtx-cu12==12.6.77 -grpcio==1.71.0 -sniffio==1.3.1 -aiosignal==1.3.2 -pillow==11.2.1 -triton==3.3.0 -greenlet==3.2.2 -orjson==3.10.18 -SQLAlchemy==2.0.41 -typing_extensions==4.13.2 -cloudpickle==3.1.1 -fastjsonschema==2.21.1 -transformers==4.52.4 -uvicorn==0.34.2 -typing-inspect==0.9.0 -pyzmq==26.4.0 -parso==0.8.4 -langchain==0.3.25 -yarl==1.20.0 -MarkupSafe==3.0.2 -mlflow-skinny==2.22.0 -tensorboard-data-server==0.7.2 -contourpy==1.3.2 -jupyter-events==0.12.0 -Mako==1.3.10 -wcwidth==0.2.13 -pyparsing==3.2.3 -Jinja2==3.1.6 -blinker==1.9.0 -json5==0.12.0 -opencv-python-headless==4.11.0.86 -ipywidgets==8.1.7 -packaging==25.0 -nbformat==5.10.4 -pytz==2025.2 -async-lru==2.0.5 -absl-py==2.2.2 -beautifulsoup4==4.13.4 -ipython==9.3.0 -executing==2.2.0 -cycler==0.12.1 -pydantic==2.11.5 -torchinfo==1.8.0 -math-verify==0.7.0 -regex==2024.11.6 -annotated-types==0.7.0 -smmap==5.0.2 diff --git a/wandb/offline-run-20250613_022053-ok4hunfv/files/wandb-metadata.json b/wandb/offline-run-20250613_022053-ok4hunfv/files/wandb-metadata.json deleted file mode 100644 index 06cfe18..0000000 --- a/wandb/offline-run-20250613_022053-ok4hunfv/files/wandb-metadata.json +++ /dev/null @@ -1,46 +0,0 @@ -{ - "os": "Linux-6.11.0-26-generic-x86_64-with-glibc2.39", - "python": "CPython 3.12.3", - "startedAt": "2025-06-12T20:50:53.785044Z", - "args": [ - "--model_id", - "google/gemma-3-4b-pt", - "--dataset_id", - "ariG23498/license-detection-paligemma", - "--batch_size", - "8", - "--lr", - "2e-5", - "--checkpoint_id", - "oreonmayo/gemma-3-4b-pt-object-detection-aug", - "--peft_type", - "lora", - "--peft_config", - "peft_configs/lora_configs.yaml" - ], - "program": "/home/manodeep/projects/NN/neural_networks_repo/finetuning/gemma-3_lora/gemma3-object-detection/finetune.py", - "codePath": "finetune.py", - "git": { - "remote": "https://github.com/Manodeepray/gemma3-object-detection", - "commit": "4d0d9cbaaaf2d548d7c42c7e4ec546c0d85e7143" - }, - "root": "/home/manodeep/projects/NN/neural_networks_repo/finetuning/gemma-3_lora/gemma3-object-detection", - "host": "mayo", - "executable": "/home/manodeep/projects/NN/.venv/bin/python", - "codePathLocal": "finetune.py", - "cpu_count": 4, - "cpu_count_logical": 8, - "disk": { - "/": { - "total": "208557645824", - "used": "87174987776" - } - }, - "memory": { - "total": "16297222144" - }, - "cpu": { - "count": 4, - "countLogical": 8 - } -} \ No newline at end of file diff --git a/wandb/offline-run-20250613_022053-ok4hunfv/logs/debug-core.log b/wandb/offline-run-20250613_022053-ok4hunfv/logs/debug-core.log deleted file mode 120000 index 592efb5..0000000 --- a/wandb/offline-run-20250613_022053-ok4hunfv/logs/debug-core.log +++ /dev/null @@ -1 +0,0 @@ -/home/manodeep/.cache/wandb/logs/core-debug-20250613_022053.log \ No newline at end of file diff --git a/wandb/offline-run-20250613_022053-ok4hunfv/logs/debug-internal.log b/wandb/offline-run-20250613_022053-ok4hunfv/logs/debug-internal.log deleted file mode 100644 index bf7a12d..0000000 --- a/wandb/offline-run-20250613_022053-ok4hunfv/logs/debug-internal.log +++ /dev/null @@ -1,8 +0,0 @@ -{"time":"2025-06-13T02:20:53.797779656+05:30","level":"INFO","msg":"stream: starting","core version":"0.20.1","symlink path":"/home/manodeep/projects/NN/neural_networks_repo/finetuning/gemma-3_lora/gemma3-object-detection/wandb/offline-run-20250613_022053-ok4hunfv/logs/debug-core.log"} -{"time":"2025-06-13T02:20:53.903803604+05:30","level":"WARN","msg":"GraphQL client is nil, skipping feature loading"} -{"time":"2025-06-13T02:20:53.904210076+05:30","level":"INFO","msg":"stream: created new stream","id":"ok4hunfv"} -{"time":"2025-06-13T02:20:53.904243526+05:30","level":"INFO","msg":"stream: started","id":"ok4hunfv"} -{"time":"2025-06-13T02:20:53.904330248+05:30","level":"INFO","msg":"sender: started","stream_id":"ok4hunfv"} -{"time":"2025-06-13T02:20:53.904303338+05:30","level":"INFO","msg":"writer: Do: started","stream_id":"ok4hunfv"} -{"time":"2025-06-13T02:20:53.904357482+05:30","level":"INFO","msg":"handler: started","stream_id":"ok4hunfv"} -{"time":"2025-06-13T02:20:53.911351481+05:30","level":"INFO","msg":"Starting system monitor"} diff --git a/wandb/offline-run-20250613_022053-ok4hunfv/logs/debug.log b/wandb/offline-run-20250613_022053-ok4hunfv/logs/debug.log deleted file mode 100644 index ccf00f8..0000000 --- a/wandb/offline-run-20250613_022053-ok4hunfv/logs/debug.log +++ /dev/null @@ -1,21 +0,0 @@ -2025-06-13 02:20:53,564 INFO MainThread:54298 [wandb_setup.py:_flush():81] Current SDK version is 0.20.1 -2025-06-13 02:20:53,564 INFO MainThread:54298 [wandb_setup.py:_flush():81] Configure stats pid to 54298 -2025-06-13 02:20:53,564 INFO MainThread:54298 [wandb_setup.py:_flush():81] Loading settings from /home/manodeep/.config/wandb/settings -2025-06-13 02:20:53,564 INFO MainThread:54298 [wandb_setup.py:_flush():81] Loading settings from /home/manodeep/projects/NN/neural_networks_repo/finetuning/gemma-3_lora/gemma3-object-detection/wandb/settings -2025-06-13 02:20:53,564 INFO MainThread:54298 [wandb_setup.py:_flush():81] Loading settings from environment variables -2025-06-13 02:20:53,564 INFO MainThread:54298 [wandb_init.py:setup_run_log_directory():703] Logging user logs to /home/manodeep/projects/NN/neural_networks_repo/finetuning/gemma-3_lora/gemma3-object-detection/wandb/offline-run-20250613_022053-ok4hunfv/logs/debug.log -2025-06-13 02:20:53,564 INFO MainThread:54298 [wandb_init.py:setup_run_log_directory():704] Logging internal logs to /home/manodeep/projects/NN/neural_networks_repo/finetuning/gemma-3_lora/gemma3-object-detection/wandb/offline-run-20250613_022053-ok4hunfv/logs/debug-internal.log -2025-06-13 02:20:53,564 INFO MainThread:54298 [wandb_init.py:init():831] calling init triggers -2025-06-13 02:20:53,564 INFO MainThread:54298 [wandb_init.py:init():836] wandb.init called with sweep_config: {} -config: {'dataset_id': 'ariG23498/license-detection-paligemma', 'model_id': 'google/gemma-3-4b-pt', 'checkpoint_id': 'sergiopaniego/gemma-3-4b-pt-object-detection-aug', 'device': 'cpu', 'dtype': torch.bfloat16, 'batch_size': 8, 'learning_rate': 2e-05, 'project_name': 'gemma_finetune', '_wandb': {}} -2025-06-13 02:20:53,564 INFO MainThread:54298 [wandb_init.py:init():872] starting backend -2025-06-13 02:20:53,775 INFO MainThread:54298 [wandb_init.py:init():875] sending inform_init request -2025-06-13 02:20:53,784 INFO MainThread:54298 [wandb_init.py:init():883] backend started and connected -2025-06-13 02:20:53,786 INFO MainThread:54298 [wandb_init.py:init():956] updated telemetry -2025-06-13 02:20:53,795 INFO MainThread:54298 [wandb_init.py:init():980] communicating run to backend with 90.0 second timeout -2025-06-13 02:20:53,907 INFO MainThread:54298 [wandb_init.py:init():1032] starting run threads in backend -2025-06-13 02:20:54,061 INFO MainThread:54298 [wandb_run.py:_console_start():2453] atexit reg -2025-06-13 02:20:54,061 INFO MainThread:54298 [wandb_run.py:_redirect():2301] redirect: wrap_raw -2025-06-13 02:20:54,061 INFO MainThread:54298 [wandb_run.py:_redirect():2370] Wrapping output streams. -2025-06-13 02:20:54,061 INFO MainThread:54298 [wandb_run.py:_redirect():2393] Redirects installed. -2025-06-13 02:20:54,062 INFO MainThread:54298 [wandb_init.py:init():1078] run started, returning control to user process diff --git a/wandb/offline-run-20250613_022053-ok4hunfv/run-ok4hunfv.wandb b/wandb/offline-run-20250613_022053-ok4hunfv/run-ok4hunfv.wandb deleted file mode 100644 index e69de29..0000000