Geohunterr commited on
Commit
0622c16
1 Parent(s): 6e31e24

Training in progress, epoch 0

Browse files
Logs/events.out.tfevents.1718373338.e29ec45d9208.134.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a106a9a8693eff401d9813a4fa8bb6098b827c256b7c5568e7b5e02cd580abc
3
+ size 5617
Logs/events.out.tfevents.1718373362.e29ec45d9208.134.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:87c12907a4a62ad6570468cc1d8f02909c96f3227c4f2e1018aa4f92c4ee28b4
3
+ size 6094
Untitled.ipynb CHANGED
@@ -260,7 +260,9 @@
260
  "cell_type": "code",
261
  "execution_count": 6,
262
  "id": "d2566bf3-1104-49b6-bd87-042552f00088",
263
- "metadata": {},
 
 
264
  "outputs": [
265
  {
266
  "name": "stdout",
@@ -299,7 +301,9 @@
299
  "cell_type": "code",
300
  "execution_count": 7,
301
  "id": "10910949-9e5d-42d9-859e-7c3692f3a945",
302
- "metadata": {},
 
 
303
  "outputs": [
304
  {
305
  "name": "stdout",
@@ -425,7 +429,7 @@
425
  {
426
  "data": {
427
  "application/vnd.jupyter.widget-view+json": {
428
- "model_id": "c690568e9e534530994ef0cd77e4acc6",
429
  "version_major": 2,
430
  "version_minor": 0
431
  },
@@ -468,8 +472,8 @@
468
  "\n",
469
  " <div>\n",
470
  " \n",
471
- " <progress value='647' max='1290' style='width:300px; height:20px; vertical-align: middle;'></progress>\n",
472
- " [ 647/1290 2:34:35 < 2:34:07, 0.07 it/s, Epoch 4.99/10]\n",
473
  " </div>\n",
474
  " <table border=\"1\" class=\"dataframe\">\n",
475
  " <thead>\n",
@@ -480,21 +484,6 @@
480
  " </tr>\n",
481
  " </thead>\n",
482
  " <tbody>\n",
483
- " <tr>\n",
484
- " <td>0</td>\n",
485
- " <td>1.434000</td>\n",
486
- " <td>0.000498</td>\n",
487
- " </tr>\n",
488
- " <tr>\n",
489
- " <td>2</td>\n",
490
- " <td>0.000100</td>\n",
491
- " <td>0.000185</td>\n",
492
- " </tr>\n",
493
- " <tr>\n",
494
- " <td>4</td>\n",
495
- " <td>0.000000</td>\n",
496
- " <td>0.000146</td>\n",
497
- " </tr>\n",
498
  " </tbody>\n",
499
  "</table><p>"
500
  ],
@@ -504,28 +493,6 @@
504
  },
505
  "metadata": {},
506
  "output_type": "display_data"
507
- },
508
- {
509
- "name": "stderr",
510
- "output_type": "stream",
511
- "text": [
512
- "/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py:1132: FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`.\n",
513
- " warnings.warn(\n",
514
- "/usr/local/lib/python3.10/dist-packages/torch/utils/checkpoint.py:31: UserWarning: None of the inputs have requires_grad=True. Gradients will be None\n",
515
- " warnings.warn(\"None of the inputs have requires_grad=True. Gradients will be None\")\n",
516
- "/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py:1132: FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`.\n",
517
- " warnings.warn(\n",
518
- "/usr/local/lib/python3.10/dist-packages/torch/utils/checkpoint.py:31: UserWarning: None of the inputs have requires_grad=True. Gradients will be None\n",
519
- " warnings.warn(\"None of the inputs have requires_grad=True. Gradients will be None\")\n",
520
- "/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py:1132: FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`.\n",
521
- " warnings.warn(\n",
522
- "/usr/local/lib/python3.10/dist-packages/torch/utils/checkpoint.py:31: UserWarning: None of the inputs have requires_grad=True. Gradients will be None\n",
523
- " warnings.warn(\"None of the inputs have requires_grad=True. Gradients will be None\")\n",
524
- "/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py:1132: FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`.\n",
525
- " warnings.warn(\n",
526
- "/usr/local/lib/python3.10/dist-packages/torch/utils/checkpoint.py:31: UserWarning: None of the inputs have requires_grad=True. Gradients will be None\n",
527
- " warnings.warn(\"None of the inputs have requires_grad=True. Gradients will be None\")\n"
528
- ]
529
  }
530
  ],
531
  "source": [
@@ -663,7 +630,15 @@
663
  "\n",
664
  "#=================3) Declare Tokenizer for Tokenizing Dataset=================================\n",
665
  "\n",
666
- "TokenToIDFn = ModelProcessor.tokenizer.convert_tokens_to_ids(\"<image>\")\n",
 
 
 
 
 
 
 
 
667
  "\n",
668
  "def TokenGeneratorFn(DatasetEntries):\n",
669
  " try:\n",
@@ -734,7 +709,7 @@
734
  "\n",
735
  " logging_dir=\"/workspace/Logs\",\n",
736
  " logging_steps=100,\n",
737
- " num_train_epochs=10,\n",
738
  " per_device_train_batch_size=16,\n",
739
  " # per_device_eval_batch_size=16,\n",
740
  " gradient_accumulation_steps=4,\n",
@@ -843,7 +818,7 @@
843
  },
844
  {
845
  "cell_type": "code",
846
- "execution_count": 1,
847
  "id": "f5855fa1-23e4-4542-a9cf-406a1308c84e",
848
  "metadata": {},
849
  "outputs": [],
 
260
  "cell_type": "code",
261
  "execution_count": 6,
262
  "id": "d2566bf3-1104-49b6-bd87-042552f00088",
263
+ "metadata": {
264
+ "scrolled": true
265
+ },
266
  "outputs": [
267
  {
268
  "name": "stdout",
 
301
  "cell_type": "code",
302
  "execution_count": 7,
303
  "id": "10910949-9e5d-42d9-859e-7c3692f3a945",
304
+ "metadata": {
305
+ "scrolled": true
306
+ },
307
  "outputs": [
308
  {
309
  "name": "stdout",
 
429
  {
430
  "data": {
431
  "application/vnd.jupyter.widget-view+json": {
432
+ "model_id": "8f49b17e69404b56ae7925565eead95c",
433
  "version_major": 2,
434
  "version_minor": 0
435
  },
 
472
  "\n",
473
  " <div>\n",
474
  " \n",
475
+ " <progress value='65' max='645' style='width:300px; height:20px; vertical-align: middle;'></progress>\n",
476
+ " [ 65/645 14:25 < 2:12:46, 0.07 it/s, Epoch 0.49/5]\n",
477
  " </div>\n",
478
  " <table border=\"1\" class=\"dataframe\">\n",
479
  " <thead>\n",
 
484
  " </tr>\n",
485
  " </thead>\n",
486
  " <tbody>\n",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
487
  " </tbody>\n",
488
  "</table><p>"
489
  ],
 
493
  },
494
  "metadata": {},
495
  "output_type": "display_data"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
496
  }
497
  ],
498
  "source": [
 
630
  "\n",
631
  "#=================3) Declare Tokenizer for Tokenizing Dataset=================================\n",
632
  "\n",
633
+ "TokenToIDFn = ModelProcessor.tokenizer.convert_tokens_to_ids(\"<image>\") \n",
634
+ "\n",
635
+ "ModelToFineTune = PaliGemmaForConditionalGeneration.from_pretrained(\n",
636
+ " ModelID,\n",
637
+ " torch_dtype=torch.bfloat16\n",
638
+ ").to(device)\n",
639
+ "\n",
640
+ "for param in ModelToFineTune.vision_tower.parameters():\n",
641
+ " \n",
642
  "\n",
643
  "def TokenGeneratorFn(DatasetEntries):\n",
644
  " try:\n",
 
709
  "\n",
710
  " logging_dir=\"/workspace/Logs\",\n",
711
  " logging_steps=100,\n",
712
+ " num_train_epochs=5,\n",
713
  " per_device_train_batch_size=16,\n",
714
  " # per_device_eval_batch_size=16,\n",
715
  " gradient_accumulation_steps=4,\n",
 
818
  },
819
  {
820
  "cell_type": "code",
821
+ "execution_count": 12,
822
  "id": "f5855fa1-23e4-4542-a9cf-406a1308c84e",
823
  "metadata": {},
824
  "outputs": [],
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:29476876c005a21c18b12182ddeca2443ce0ba63a8fb22e8c5a806a76077f831
3
  size 45258384
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f034f6296c891e2f9b7f4caf02dc78746ebc7d2855ca693f0a29142230e40b4
3
  size 45258384
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dc1c2b63389e9da8a1d8d74ce00c3e542c3bd9f4f4ba9794be2536355b834e35
3
  size 4603
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:efe5a28664d9588f5ff9737fa401845de69b7ed8777a2095966d9d6e7fb2bbae
3
  size 4603