-
Notifications
You must be signed in to change notification settings - Fork 0
/
mutational_utils.py
2181 lines (1839 loc) · 98.1 KB
/
mutational_utils.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
import numpy as np
import pandas as pd
import os
from math import floor
from itertools import product, chain
from random import shuffle, sample, choices, random
from tqdm import tqdm
import matplotlib.pyplot as plt
from Bio import SeqIO, Seq
from arnie.bpps import bpps
from arnie.utils import convert_dotbracket_to_bp_list
from Levenshtein import distance as edit_distance
from pybktree import BKTree
'''
Example construct:
26 nt. 5' fixed sequence
240 nt. Region of interest
20 nt. 8 bp barcode with UUCG tetraloop (65,536 possible barcodes)
1 nt. Single A spacer
20 nt. 3' fixed sequence
'''
BASES = ['A', 'C', 'G', 'T']
SEQ5 = 'GGGAACGACTCGAGTAGAGTCGAAAA'
SEQ3 = 'AAAAGAAACAACAACAACAAC' # with the single A spacer added already
TETRALOOP = 'TTCG'
'''
Example procedure:
for each WT, get single mutants and souble mutants of interaction
generate pad to make all equal length
each sequence add 5', random 8bp barcode, 3'
Example structure checks:
'''
MAXPROB_NONINTERACT = 0.09
MINPROB_PAIRED = 0.75
MINAVGPROB_PAIRED = 0.85
MINPROB_UNPAIRED = 0.75
MINAVGPROB_UNPAIRED = 0.85
###############################################################################
# fasta file manipulations
###############################################################################
def combine_fastas(fastas, out_fasta):
'''
Given list of fastas, appends all and saves new fasta.
'''
all_seqs = []
for fasta in fastas:
all_seqs.extend(list(SeqIO.parse(fasta, "fasta")))
print(f'Combined and saved {fastas} to {out_fasta}')
SeqIO.write(all_seqs, out_fasta, "fasta")
def format_fasta_for_submission(fasta, out_file, file_format='twist'):
'''
From a fasta, save sequences in a format for library submission
For agilent and twist: csv
For custom_array: txt with each line a new sequence
'''
if file_format == 'twist' or file_format == 'agilent':
if out_file[-4:] != '.csv':
print('twist requires .csv files, change out_file')
names = []
seqs = []
for seq in SeqIO.parse(fasta, "fasta"):
names.append(seq.name)
seqs.append(_get_dna_from_SeqRecord(seq))
df = pd.DataFrame(np.array([names, seqs]).T,
columns=["name", "sequence"])
df.to_csv(out_file, index=False)
print(f'Written {out_file} for submission to twist or agilent.')
elif file_format == 'custom_array':
if out_file[-4:] != '.txt':
print('custom_array requires .txt files, change out_file')
with open(out_file, 'w') as f:
for seq in SeqIO.parse(fasta, "fasta"):
f.write(_get_dna_from_SeqRecord(seq))
f.write('\n')
print(f'Written {out_file} for submission to custom_array.')
else:
print('file_format not supported, available: custom_array twist agilent')
def randomly_select_seqs(fasta, reject_file, prop):
'''
Given a fasta randomly select N sequences and save to new fasta
'''
all_seqs = list(SeqIO.parse(fasta, "fasta"))
N = round(len(all_seqs)*prop)
if len(all_seqs) > N:
seq_indices = list(range(len(all_seqs)))
shuffle(seq_indices)
pass_ind = seq_indices[:N]
rejected_ind = seq_indices[N:]
pass_ind.sort()
rejected_ind.sort()
pass_seqs = [seq for i, seq in enumerate(all_seqs) if i in pass_ind]
rejected_seqs = [seq for i, seq in enumerate(
all_seqs) if i in rejected_ind]
SeqIO.write(rejected_seqs, reject_file, "fasta")
SeqIO.write(pass_seqs, fasta, "fasta")
print(f'Written {fasta} for with {N} selected sequences and {reject_file} with {len(all_seqs)-N}.')
else:
SeqIO.write(all_seqs, out_fasta, "fasta")
print(f'WARNING: Written {fasta} but had less than {N} sequences so none removed.')
def get_same_length(fasta):
'''
Checks if all sequences in a given fasta are the same length
'''
seqs = list(SeqIO.parse(fasta, "fasta"))
return _get_same_length(seqs)
def check_sequences_contents(fasta,seq5=SEQ5, seq3=SEQ3,bases=BASES):
problems = []
names = []
sequences = []
seqs = list(SeqIO.parse(fasta, "fasta"))
for seq_rec in seqs:
seq = _get_dna_from_SeqRecord(seq_rec)
if seq in sequences:
problems.append(f"{seq_rec.name} is a repeated sequence.")
if seq_rec.name in names:
problems.append(f"{seq_rec.name} is a repeated name.")
if seq[:len(seq5)] != seq5:
problems.append(f"{seq_rec.name} has incorrect 5' sequence.")
if seq[-len(seq3):] != seq3:
problems.append(f"{seq_rec.name} has incorrect 3' sequence.")
for n in seq:
if n not in bases:
problems.append(f"{seq_rec.name} has non {bases} base.")
sequences.append(seq)
names.append(seq_rec.name)
return len(problems)==0, problems
def remove_seqs_already_in_other_file(fasta, other_fasta, out_file):
'''
Given 2 fasta files, removes any sequence in the first that
has same name as a sequence in the second, save non-duplicated
sequences of the first.
DOES NOT check seq itself and does not change the second fasta!
'''
all_seqs = list(SeqIO.parse(fasta, "fasta"))
other_seqs = list(SeqIO.parse(other_fasta, "fasta"))
good_seqs = _remove_seqs_in_other_list(all_seqs, other_seqs)
SeqIO.write(good_seqs, out_file, "fasta")
def get_used_barcodes(fasta, start, end):
'''
from a fasta file return all sequences between start and end inclusive
'''
all_seqs = list(SeqIO.parse(fasta, "fasta"))
barcodes = []
for record in all_seqs:
seq = _get_dna_from_SeqRecord(record)
# end is inclusive
barcode = seq[start:end+1]
barcodes.append(str(barcode))
return barcodes
###############################################################################
# get desired sequences
###############################################################################
def _fill_in_any_incomplete(seq,seqs):
incomplete_seq = {'N':['A','C','T','G'],
'R':['A','G'],
'Y':['C','T'],
'S':['C','G'],
'W':['A','T'],
'K':['T','G'],
'M':['A','C'],
'B':['C','T','G'],
'D':['A','T','G'],
'H':['A','C','T'],
'V':['A','C','G'],}
if seq == '':
return seqs
elif seq[0] in incomplete_seq:
new_seqs = []
for n in incomplete_seq[seq[0]]:
potential_seq = n+seq[1:]
potential_seqs = [s[:-len(potential_seq)]+potential_seq for s in seqs]
new_seqs.extend(potential_seqs)
return _fill_in_any_incomplete(seq[1:],new_seqs)
else:
return _fill_in_any_incomplete(seq[1:],seqs)
def get_windows(fasta, window_length, window_slide, out_fasta=None,
circularize=False, fraction_use=1, reverse_complement=False):
'''
Get sliding windows from an inputted fasta file.
Args:
fasta (str): fasta file containing sequence to get windows of
window_length (int): length (in number nt) of desired windows
window_slide (int): number nucleotide to slide between windows
out_fasta (str): if specified, save windows to fasta (default None)
circularize (bool): whether to circularize the sequence thus once it
reaches the end, create windows with 3'-5' connected (default False)
fraction_use (float): proportion (from start) of genome to use,
defaults all (default 1)
reverse_complement (bool): instead of sequence in fasta, use
reverse complement
Returns:
list of SeqRecord with the windows
if out_fasta specified, also saves these to fasta file
naming convention is the seqname_start-end with start and
end inclusive and indexing from 0.
'''
print(f'Getting all sliding windows, {window_length}nt every {window_slide}nt.')
# get sequences and initialize
seqs = list(SeqIO.parse(fasta, "fasta"))
windows = []
unused_windows = []
for seq_rec in seqs:
# loop through sequence every window_slide nucleotides
seq = _get_dna_from_SeqRecord(seq_rec)
window_limit = floor(fraction_use*len(seq))
for i in range(0, len(seq), window_slide):
# when we hit the end of sequence
if i+window_length > len(seq):
# add last window
if not circularize:
a, b = len(seq)-window_length, len(seq)
new_seq = seq[a:b]
namenum = f'{a}-{b-1}'
if reverse_complement:
new_rec = SeqIO.SeqRecord(Seq.Seq(_get_reverse_complement(new_seq)),
f'{name}_rc', '', '')
else:
new_rec = SeqIO.SeqRecord(Seq.Seq(new_seq),
f'{seq_rec.name}_{namenum}',
'', '')
if fraction_use != 1 and ((i + window_length-1) > window_limit):
unused_windows.append(new_rec)
else:
windows.append(new_rec)
break
# or circularize and add from 5' until very end
else:
a, b, c = i, len(seq), window_length-len(seq)+i
new_seq = seq[a:b]+seq[:c]
namenum = f'{a}-{c-1}'
# otherwise just add window as normal
else:
a, b = i, i+window_length
new_seq = seq[a:b]
namenum = f'{a}-{b-1}'
new_seqs = _fill_in_any_incomplete(new_seq,[new_seq])
for j,new_seq in enumerate(new_seqs):
if len(new_seqs) == 1:
name = f'{seq_rec.name}_{namenum}'
else:
name = f'{seq_rec.name}_amb{j}_{namenum}'
# save with name inclusive!
if reverse_complement:
new_rec = SeqIO.SeqRecord(Seq.Seq(_get_reverse_complement(new_seq)),
f'{name}_rc', '', '')
else:
new_rec = SeqIO.SeqRecord(Seq.Seq(new_seq),
f'{seq_rec.name}_{namenum}',
'', '')
if fraction_use != 1 and ((i + window_length-1) > window_limit):
unused_windows.append(new_rec)
else:
windows.append(new_rec)
# remove and save fraction unused
if fraction_use != 1:
unused_file = f'{out_fasta.rsplit(".",1)[0]}_unused.{out_fasta.rsplit(".",1)[1]}'
SeqIO.write(unused_windows, unused_file, "fasta")
print(f'Saved unused windows to {unused_file}.')
# save file
if out_fasta is not None:
SeqIO.write(windows, out_fasta, "fasta")
print(f'Saved windows to {out_fasta}.')
# return list of windows
return windows
def get_all_single_mutants(fasta, out_fasta=None, bases=BASES):
'''
Get all single mutants from sequences in a fasta file.
Args:
fasta (str): fasta file containing sequence to get mutants of
out_fasta (str): if specified, save mutants to fasta (default None)
bases (list): what bases can be selected as mutants (default ['A', 'C', 'G', 'T'])
Returns:
list of SeqRecord with the mutants
if out_fasta specified, also saves these to fasta file
naming convention is the seqname_#wt-mutant eg for sequence P4P6
mutant with 138 mutated from G to A is: P4P6_138G-A
Indexing from 0.
Generally, total is 3*length seq
'''
print("Getting all single mutants.")
# get all sequences and initialize
all_WT = SeqIO.parse(fasta, "fasta")
all_single_mutants = []
for record in all_WT:
seq = _get_dna_from_SeqRecord(record)
# at each position, get single mutants
for i in range(len(seq)):
for mut in bases:
if mut != seq[i]:
name = f' {record.id}_{i}{seq[i]}-{mut}'
new_seq = seq[:i]+mut+seq[i+1:]
new_mut = SeqIO.SeqRecord(Seq.Seq(new_seq), name, '', '')
all_single_mutants.append(new_mut)
# save file
if out_fasta is not None:
SeqIO.write(all_single_mutants, out_fasta, "fasta")
print(f'Saved single mutants to {out_fasta}.')
return all_single_mutants
def get_all_double_mutants(fasta, regionAs, regionBs,
out_fasta=None, bases=BASES):
'''
Get all double mutants between specified region for all sequences in a fasta file.
Args:
fasta (str): fasta file containing sequence to get mutants of
regionAs (list of lists of int): for each sequence, a list of indices specifying first region
regionBs (list of lists of int): for each sequence, a list of indices specifying second region
out_fasta (str): if specified, save mutants to fasta (default None)
bases (list): what bases can be selected as mutants (default ['A', 'C', 'G', 'T'])
Returns:
list of SeqRecord with the mutants
if out_fasta specified, also saves these to fasta file
naming convention is the seqname_#wt-mutant eg for sequence P4P6
mutant with 138 mutated from G to A and 150 C to T is: P4P6_138G-A_150C-T
Indexing from 0 on mutants are ordered by nucleotide number.
Generally, total is 9*lengthA*lengthB
'''
print("Getting all double mutants.")
# get all sequences and initialize
all_WT = list(SeqIO.parse(fasta, "fasta"))
all_double_mutants = []
# print(regionAs,regionBs)
# check user specified regions for every sequence in the fasta
if len(all_WT) != len(regionAs) or len(all_WT) != len(regionBs):
print(f'WARNING: you must list regions (as a list of lists) for all sequences in the fasta, number sequences in fasta {len(all_WT)} and number of regions specified {len(regionAs)} {len(regionBs)}')
for record, regionA_s, regionB_s in zip(all_WT, regionAs, regionBs):
seq = _get_dna_from_SeqRecord(record)
for regionA,regionB in zip(regionA_s,regionB_s):
# at each position pair, get double mutants
for i in regionA:
for mutA in bases:
if mutA != seq[i]:
for j in regionB:
for mutB in bases:
if mutB != seq[j]:
# get mutant
# name according to convention, in order, index at 1
if i == j:
continue
elif i < j:
new_seq = seq[:i]+mutA + \
seq[i+1:j]+mutB+seq[j+1:]
name = f' {record.id}_{i}{seq[i]}-{mutA}_{j}{seq[j]}-{mutB}'
else:
new_seq = seq[:j]+mutB + \
seq[j+1:i]+mutA+seq[i+1:]
name = f' {record.id}_{j}{seq[j]}-{mutB}_{i}{seq[i]}-{mutA}'
new_mut = SeqIO.SeqRecord(
Seq.Seq(new_seq), name, '', '')
all_double_mutants.append(new_mut)
# save file
if out_fasta is not None:
SeqIO.write(all_double_mutants, out_fasta, "fasta")
print(f'Saved all double mutants between the 2 regions to {out_fasta}.')
return all_double_mutants
def get_wcf_rescue_mutants(fasta, bp_sets, out_fasta=None,
wfc_base_pairs=['AT', 'TA', 'CG', 'GC']):
'''
Get all Watson-Crick-Franklin rescue mutants between specified base-pairs for all sequences in a fasta file.
Args:
fasta (str): fasta file containing sequence to get mutants of
bp_sets (list of lists of pairs): for each sequence, a list of base-pairs as a tuple or list of int
out_fasta (str): if specified, save mutants to fasta (default None)
wfc_base_pairs (list): list of potential base pairs to rescue with
(default ['AT', 'TA', 'CG', 'GC'])
Returns:
list of SeqRecord with the mutants
if out_fasta specified, also saves these to fasta file
naming convention is the seqname_#wt-mutant eg for sequence P4P6
mutant with 138 mutated from G to A and 150 C to T is: P4P6_138G-A_150C-T
Indexing from 0 on mutants are ordered by nucleotide number.
Generally, total is 3*numbps
'''
print("Getting all rescue mutants.")
# get all sequences and initialize
all_WT = list(SeqIO.parse(fasta, "fasta"))
all_rescue_mutants = []
# check user specified set of basepairs for every sequence in the fasta
if len(all_WT) != len(bp_sets):
print(f'WARNING: bps must be a list, one for each sequence in fasta, of lists of basepairs to rescue for that sequence. You have {len(all_WT)} inputted sequences and {len(bp_sets)} base-pair sets.')
for record, bps in zip(all_WT, bp_sets):
seq = _get_dna_from_SeqRecord(record)
# at each base pair, get rescue mutants
for bp in bps:
current_bp = seq[bp[0]]+seq[bp[1]]
for new_bp in wfc_base_pairs:
if new_bp != current_bp:
# get mutant
# name according to convention, in order, index at 1
if bp[0] < bp[1]:
name = f' {record.id}_{bp[0]}{seq[bp[0]]}-{new_bp[0]}_{bp[1]}{seq[bp[1]]}-{new_bp[1]}'
new_seq = seq[:bp[0]]+new_bp[0] + \
seq[bp[0]+1:bp[1]]+new_bp[1]+seq[bp[1]+1:]
else:
name = f' {record.id}_{bp[1]}{seq[bp[1]]}-{new_bp[1]}_{bp[0]}{seq[bp[0]]}-{new_bp[0]}'
new_seq = seq[:bp[1]]+new_bp[1] + \
seq[bp[1]+1:bp[0]]+new_bp[0]+seq[bp[0]+1:]
new_mut = SeqIO.SeqRecord(
Seq.Seq(new_seq), name, '', '')
all_rescue_mutants.append(new_mut)
# save file
if out_fasta is not None:
SeqIO.write(all_rescue_mutants, out_fasta, "fasta")
print(f'Saved all rescue mutants to {out_fasta}.')
return all_rescue_mutants
###############################################################################
# add library parts
###############################################################################
def add_known_pads(fasta, pad5_dict, pad3_dict, out_fasta=None):
'''
From fasta prepend and append given sequences
Args:
fasta (str): fasta file containing sequence to get mutants of
pad5_dict (dict int:str): sequence length : pad to add on 5'
pad3_dict (dict int:str): sequence length : pad to add on 3'
out_fasta (str): if specified, save mutants to fasta (default None)
Returns:
list of SeqRecord with pads added
# pad# where numbers are length of pad
if out_fasta specified, also saves these to fasta file naming
'''
# get all sequences and initialize
all_WT = list(SeqIO.parse(fasta, "fasta"))
all_seqs = []
for record in all_WT:
seq = _get_dna_from_SeqRecord(record)
# get pad for this length
pad5 = pad5_dict[len(seq)]
pad3 = pad3_dict[len(seq)]
name = f' {record.id}_{len(pad5)}pad{len(pad3)}'
new_seq = SeqIO.SeqRecord(Seq.Seq(pad5+seq+pad3), name, '', '')
all_seqs.append(new_seq)
# save file
if out_fasta is not None:
SeqIO.write(all_seqs, out_fasta, "fasta")
print(f'Saved all with correct constant pad added to {out_fasta}.')
return all_seqs
def get_all_barcodes(out_fasta=None, num_bp=8, num5hang=0, num3hang=0,
polyA5=0, polyA3=0,
loop=TETRALOOP, bases=BASES):
'''
Return all barcodes of specified structure
Args:
out_fasta (str): if specified, save mutants to fasta (default None)
num_bp (int): length of stem (default 8)
num5hang (int): length of random 5' hang (default 0)
num3hang (int): length of random 3' hang (default 0)
polyA5 (int): length of polyA 5' hang (placed before random) (default 0)
polyA3 (int): length of polyA 3' hang (placed after random) (default 0)
loop (str): sequence of loop
bases (list): list of bases that can be used
Returns:
list of SeqRecord of all possible barcodes
if out_fasta specified, also saves these to fasta file
'''
print("Getting all possible barcodes.")
all_barcodes = []
# get all possible combinations of bases for random/barcode regions
for x in product(bases, repeat=num5hang+num_bp+num3hang):
uid = ''.join(x)
# split barcode in stem and hang regions
hang5 = uid[:num5hang]
if num3hang == 0:
hang3 = ''
stemA = uid[num5hang:]
else:
hang3 = uid[-num3hang:]
stemA = uid[num5hang:-num3hang]
stemB = _get_reverse_complement(stemA)
# put all barcode parts together
seq = ("A"*polyA5)+hang5+stemA+loop+stemB+hang3+("A"*polyA3)
name = f' stem{stemA}_{hang5}hang{hang3}_{polyA5}polyA{polyA3}'
seq_rec = SeqIO.SeqRecord(Seq.Seq(seq), name, '', '')
all_barcodes.append(seq_rec)
# save
if out_fasta is not None:
SeqIO.write(all_barcodes, out_fasta, "fasta")
print(f'Saved all barcodes to {out_fasta}.')
return all_barcodes
def add_pad(fasta, out_fasta, bases=BASES, share_pad='same_length',
epsilon_punpaired=MINPROB_UNPAIRED,
epsilon_interaction=MAXPROB_NONINTERACT,
epsilon_avg_punpaired=MINAVGPROB_UNPAIRED,
epsilon_paired=MINPROB_PAIRED,
epsilon_avg_paired=MINAVGPROB_PAIRED,
loop=TETRALOOP, hang=3, polyAhang=0, min_num_samples=30,
max_prop_bad=0.05, pad_side='both',
min_length_stem=4, max_length_stem=12,
num_pads_reduce=100, percent_reduce_prob=10,pad_to_length=None):
'''
Given a fasta of sequence pad all sequence to the same length
Args:
fasta (str): fasta file containing sequence to get mutants of
out_fasta (str): if specified, save mutants to fasta (default None)
share_pad (str): either all different pads (none),
all same for sequences of same length (same_length),
all same where sequence of different length are just truncated (all)
epsilon_interaction (float): Maximum base-pair-probability for 2 regions to be considered non-interacting
epsilon_punpaired (float): Minimum probability unpaired for region to be unpaired
epsilon_avg_punpaired (float): Average probability unpaired for region to be unpaired
epsilon_paired (float): Minimum base-pair-probability for 2 regions to be considered paired
epsilon_avg_paired (float): Average base-pair-probability for 2 regions to be considered paired
loop (str): constant sequence for loop of stem-loop
hang (int): distance between sequence and the stem will be this or this+1, random sequence (default 3)
polyAhang (int): distance between sequence and the stem will be this or this+1, polyA (default 3)
min_length_stem (int): minimum number base-pairs to form stem
max_length_stem (int): maximum number of base-pairs to form stem
min_num_samples (int): minimum number of samples to test pad structure (default 30)
max_prop_bad (float): of the samples sequences, proportion that can be bad (default 0.05)
pad_side (str): whether split pad ('both') or force to be exclusively 5' or 3'
num_pads_reduce (int): number of pads to try before reducing probability (default 100)
percent_reduce_prob (float): percent to reduce probabilities each time (default 10)
Returns:
list of SeqRecord with pads added
# pad# where numbers are length of pad
if out_fasta specified, also saves these to fasta file naming added _
'''
print('WARNING this code will likely be obsoletely by a code that takes other regions into account when choosing the pad.')
# get sequences and sort by length
seqs = list(SeqIO.parse(fasta, "fasta"))
seq_by_length = {}
for seq_rec in seqs:
len_seq = len(seq_rec.seq)
if len_seq in seq_by_length:
seq_by_length[len_seq].append(seq_rec)
else:
seq_by_length[len_seq] = [seq_rec]
# for each length, randomly select 1% of sequences or min_num_samples
selected_sec = []
max_bad_structs = {}
for len_seq, seq_group in seq_by_length.items():
number_to_select = min(len(seq_group), max(
min_num_samples, len(seq_group)*0.01))
number_to_select = round(number_to_select)
selected_sec.append(sample(seq_group, k=number_to_select))
max_bad_structs[len_seq] = floor(max_prop_bad*number_to_select)
print(f'Pad for length {len_seq} search using {number_to_select} sequences for structure check for each length.')
# get max length of pad needed
if pad_to_length is None:
desired_len = max(seq_by_length.keys())
else:
desired_len = pad_to_length
porp_reduce = (1-(percent_reduce_prob/100))
# if want all pads to be random
if share_pad == 'none':
print("Getting random pads for each sequence, not guaranteed to be unique.")
padded_seqs = []
for seq_length, seqs in seq_by_length.items():
# for each length get the structure of the 3' and 5' pad
pad_length = desired_len-seq_length
structs, regions = _get_5_3_split(pad_length, hang, polyAhang,
min_length_stem, max_length_stem,
pad_side, loop, seq_length)
# for each sequence search for good_pad
for seq in list(seqs):
good_pad = False
while not good_pad:
# get random pad
pad5 = _get_random_barcode(num_bp=structs["5"]['bp'],
num3hang=structs["5"]['hang_rand'],
polyA3=structs["5"]['hang_polyA'],
loop=structs["5"]['loop'])
pad3 = _get_random_barcode(num_bp=structs["3"]['bp'],
num5hang=structs["3"]['hang_rand'],
polyA5=structs["3"]['hang_polyA'],
loop=structs["3"]['loop'])
pad5 = _get_dna_from_SeqRecord(pad5)
pad3 = _get_dna_from_SeqRecord(pad3)
# get full sequence and check structure
full_seq = pad5+_get_dna_from_SeqRecord(seq)+pad3
full_seq_name = f'{seq.name}_{len(pad5)}pad{len(pad3)}'
struct_results = check_struct_bpp(full_seq,
regions['unpaired'], regions['pairedA'],
regions['pairedB'], regions['noninteractA'], regions['noninteractB'],
epsilon_interaction=epsilon_interaction,
epsilon_punpaired=epsilon_punpaired,
epsilon_avg_punpaired=epsilon_avg_punpaired,
epsilon_paired=epsilon_paired,
epsilon_avg_paired=epsilon_avg_paired)
good_pad = struct_results["pass"]
# if pass add final sequence
padded_seq = SeqIO.SeqRecord(Seq.Seq(full_seq),
full_seq_name, '', '')
padded_seqs.append(padded_seq)
# if want sequences of same length to share same pad
elif share_pad == 'same_length':
pads_by_len = {}
for seqs in selected_sec:
# get length of pad for this group, if 0 done
len_group = len(str(seqs[0].seq))
length_to_add = desired_len-len_group
if length_to_add == 0:
pads_by_len[len_group] = ['', '']
continue
print(f"Searching for pad for group len {len_group}")
# split length of pad to either end of the sequence and get regions
structs, regions = _get_5_3_split(length_to_add, hang, polyAhang,
min_length_stem, max_length_stem,
pad_side, loop, len_group)
print(f"Finding a {structs['5']['N']}nt 5' pad with {structs['5']['bp']}bp stem {structs['5']['hang_rand']}nt random hang {structs['5']['hang_polyA']}nt polyA hang.")
print(f"Finding a {structs['3']['N']}nt 3' pad with {structs['3']['bp']}bp stem {structs['3']['hang_rand']}nt random hang {structs['3']['hang_polyA']}nt polyA hang.")
# loop through to find pad that works
good_pad = False
seq_count = {'unpaired': 0, 'paired': 0, 'interaction': 0}
# current_pad = 0
while not good_pad:
# if tried enough relax the probability contstraints
prob_factor = {}
for type_error, count in seq_count.items():
mutiplier = (count // num_pads_reduce)
prob_factor[type_error] = porp_reduce**mutiplier
if (count-mutiplier) % num_pads_reduce == 0 and count != 0:
seq_count[type_error] += 1
print(f'For {len_group}, failed to find pad from {count-mutiplier} pads because of {type_error}, reducing probabilities needed by a factor of {prob_factor[type_error]}.')
# get random pad
pad5 = _get_random_barcode(num_bp=structs["5"]['bp'],
num3hang=structs["5"]['hang_rand'],
polyA3=structs["5"]['hang_polyA'],
loop=structs["5"]['loop'])
pad3 = _get_random_barcode(num_bp=structs["3"]['bp'],
num5hang=structs["3"]['hang_rand'],
polyA5=structs["3"]['hang_polyA'],
loop=structs["3"]['loop'])
# chek all samples sequences
bad_count = 0
for i, seq in enumerate(seqs):
# if i % 50 == 0 and i != 0:
# print(i)
# get full sequence and check its structure
full_seq = pad5 + _get_dna_from_SeqRecord(seq) + pad3
struct_results = check_struct_bpp(full_seq,
regions['unpaired'], regions['pairedA'],
regions['pairedB'], regions['noninteractA'], regions['noninteractB'],
epsilon_interaction=epsilon_interaction,
epsilon_punpaired=epsilon_punpaired,
epsilon_avg_punpaired=epsilon_avg_punpaired,
epsilon_paired=epsilon_paired,
epsilon_avg_paired=epsilon_avg_paired,
prob_factor=prob_factor)
# if not good structure add to count
# stop if reached maximal bad
good_pad = struct_results["pass"]
if not good_pad:
seq_count = _update_seq_count(seq_count,struct_results)
bad_count += 1
if bad_count >= max_bad_structs[len_group]:
break
pads_by_len[len_group] = [_get_dna_from_SeqRecord(
pad5), _get_dna_from_SeqRecord(pad3)]
print('Found pads, now adding pads.')
padded_seqs = []
for seq_length, seqs in seq_by_length.items():
pad5, pad3 = pads_by_len[seq_length]
for seq in list(seqs):
full_seq = pad5+_get_dna_from_SeqRecord(seq)+pad3
full_seq_name = f'{seq.name}_{len(pad5)}pad{len(pad3)}'
padded_seq = SeqIO.SeqRecord(Seq.Seq(full_seq),
full_seq_name, '', '')
padded_seqs.append(padded_seq)
# if want all sequences to share same pad (truncated as needed)
elif share_pad == 'all':
print(f"Searching for pad for group all sequences")
print('WARNING depending on your set of sequences and structure desired this may be a near impossible random search space!')
# get all pad structures for the pad to be shared across all
pad_lengths = [desired_len-x for x in seq_by_length.keys()]
all_structs, cutoff_by_len, regions_by_len = _get_5_3_split_multi(pad_lengths, hang, polyAhang,
min_length_stem, max_length_stem,
pad_side, loop, list(seq_by_length.keys()))
# loop until find a good pad
good_pad = False
seq_count = {'unpaired': 0, 'paired': 0, 'interaction': 0}
while not good_pad:
# get random pad
pad5_full = ''
pad3_full = ''
for structs in all_structs:
pad5_full = pad5_full+_get_random_barcode(num_bp=structs["5"]['bp'],
num3hang=structs["5"]['hang_rand'],
polyA3=structs["5"]['hang_polyA'],
loop=structs["5"]['loop'])
pad3_full = _get_random_barcode(num_bp=structs["3"]['bp'],
num5hang=structs["3"]['hang_rand'],
polyA5=structs["3"]['hang_polyA'],
loop=structs["3"]['loop']) + pad3_full
# for all sequences
for seqs in selected_sec:
bad_count = 0
# if tried enough relax the probability constraints
prob_factor = {}
for type_error, count in seq_count.items():
mutiplier = (count // num_pads_reduce)
prob_factor[type_error] = porp_reduce**mutiplier
if (count-mutiplier) % num_pads_reduce == 0 and count != 0:
seq_count[type_error] += 1
print(f'For {len_group}, failed to find pad from {count-mutiplier} pads because of {type_error}, reducing probabilities needed by a factor of {prob_factor[type_error]}.')
# for each length get regions and cutoff of full pads
len_group = len(seqs[0].seq)
regions = regions_by_len[len_group]
cutoff = cutoff_by_len[len_group]
pad5 = _get_dna_from_SeqRecord(pad5_full[:cutoff[0]])
if cutoff[1] == 0:
pad3 = ''
else:
pad3 = _get_dna_from_SeqRecord(pad3_full[-cutoff[1]:])
for i, seq in enumerate(seqs):
# if i%10==0 and i!=0:
# print(i)
if len_group != desired_len:
# get full sequence and check its structure
full_seq = pad5 + _get_dna_from_SeqRecord(seq) + pad3
struct_results = check_struct_bpp(full_seq, regions['unpaired'], regions['pairedA'],
regions['pairedB'], regions['noninteractA'], regions['noninteractB'],
epsilon_interaction=epsilon_interaction,
epsilon_punpaired=epsilon_punpaired,
epsilon_avg_punpaired=epsilon_avg_punpaired,
epsilon_paired=epsilon_paired,
epsilon_avg_paired=epsilon_avg_paired,
prob_factor=prob_factor)
# if not good structure add to count
# stop if reached maximal bad
good_pad = struct_results["pass"]
if not good_pad:
seq_count = _update_seq_count(seq_count,struct_results)
bad_count += 1
if bad_count >= max_bad_structs[len_group]:
break
# add pads to the sequences
padded_seqs = []
for seq_rec in seqs:
len_group = len(seq_rec.seq)
cutoff = cutoff_by_len[len_group]
if length_to_add != 0:
pad5 = _get_dna_from_SeqRecord(pad5_full[:cutoff[0]])
if cutoff[1] == 0:
pad3 = ''
else:
pad3 = _get_dna_from_SeqRecord(pad3_full[-cutoff[1]:])
seq = _get_dna_from_SeqRecord(seq_rec)
full_seq = pad5 + seq + pad3
name = f'{seq_rec.name}_{len(pad5)}pad{len(pad3)}'
padded_seqs.append(SeqIO.SeqRecord(
Seq.Seq(full_seq), name, '', ''))
else:
padded_seqs.append(seq_rec)
else:
print("ERROR share_pad option not recognized.")
# save
if out_fasta is not None:
SeqIO.write(padded_seqs, out_fasta, "fasta")
print(f'Saved all padded sequences to {out_fasta}.')
return padded_seqs
def get_edit_distance(barcodeA, barcodeB, num_bp=0, len_loop=0):
if num_bp == 0:
dist = edit_distance(barcodeA, barcodeB)
else:
start_stem = -len_loop-(2*num_bp)
end_stem = -len_loop-num_bp
stemA = barcodeA[start_stem:end_stem]
stemB = barcodeB[start_stem:end_stem]
otherA = barcodeA[:start_stem]+barcodeB[end_stem:-num_bp]
otherB = barcodeB[:start_stem]+barcodeB[end_stem:-num_bp]
dist = 2*edit_distance(stemA, stemB)
dist += edit_distance(otherA, otherB)
return dist
def add_fixed_seq_and_barcode(fasta, out_fasta=None, seq5=SEQ5, seq3=SEQ3,
num_bp=8, num5hang=0, num5polyA=4,
loop=TETRALOOP,
epsilon_interaction=MAXPROB_NONINTERACT,
epsilon_punpaired=MINPROB_UNPAIRED,
epsilon_avg_punpaired=MINAVGPROB_UNPAIRED,
epsilon_paired=MINPROB_PAIRED,
epsilon_avg_paired=MINAVGPROB_PAIRED,
save_image_folder=None, save_bpp_fig=0,
punpaired_chunk_size=500, used_barcodes=[],
num_barcode_before_reduce=100,
percent_reduce_prob=10, min_edit=2):
'''
From a fasta of sequences, add constant regions and barcodes
Args:
fasta (str): fasta file containing sequence to get mutants of
out_fasta (str): if specified, save mutants to fasta (default None)
seq5 (str): the constant sequence to place at 5' end of all sequences
seq3 (str): the constant sequence to place at 3' end of all sequences
num_bp (int): number of base pairs to have in barcode stem (default 8)
num5hang (int): number of random nucleotides to put before the stem (default 0)
num5polyA (int): number of A to put before the barcode (stem and 5 random hang if applicable) (default 4)
loop (str): sequence of loop to have in the hairpin
epsilon_interaction (float): Maximum base-pair-probability for 2 regions to be considered non-interacting
epsilon_punpaired (float): Minimum probability unpaired for region to be unpaired
epsilon_avg_punpaired (float): Average probability unpaired for region to be unpaired
epsilon_paired (float): Minimum base-pair-probability for 2 regions to be considered paired
epsilon_avg_paired (float): Average base-pair-probability for 2 regions to be considered paired
save_image_folder (str): folder to save images to
save_bpp_fig (float): proportion of sequences to save base-pair-probability matrix figure, 0 is none, 1 is all
punpaired_chunk_size (int): max number of sequences to plot on each p unpaired plot (default 500)
used_barcodes (list): list of sequences not to use as barcodes (default [])
num_barcode_before_reduce (int): for each sequence, the number of barcodes to try
before reducing the probability thresholds by 10% (default 100)
percent_reduce_prob (float): percent to reduce probability threshold each time (default 10)
min_edit (int): minimum edit distance of barcodes, base-pairs and num5hang included (default 2)
Returns:
list of SeqRecord which are library ready
if out_fasta specified, also saves these to fasta file, _libraryready appended to names
'''
# if image folder does not exist create it
if save_image_folder is not None:
if not os.path.exists(save_image_folder):
print(f'{save_image_folder} did not exists, creating.')
os.makedirs(save_image_folder)
# get and randomly shuffle all potential barcodes
all_uids = get_all_barcodes(num_bp=num_bp, loop=loop, num5hang=num5hang,
polyA5=num5polyA)
if len(used_barcodes) != 0:
if len(all_uids[0]) != len(used_barcodes[0]):
print('ERROR: used barcodes are not the correct length')
shuffle(all_uids)
# read sequences, check all same length
seqs = list(SeqIO.parse(fasta, "fasta"))
same_length, seq_len = _get_same_length(seqs)
if not same_length:
print("ERROR: sequences not all same length.")
# initialize variables
all_full_seqs, p_unpaireds, rejected_uids = [], {}, []
pad_lines, new_lines, seqs_for_labeling, muts = [], [], [], []
current_uid, chunk_count, image_count = 0, 0, 0
seq5 = _get_dna_from_SeqRecord(seq5)
seq3 = _get_dna_from_SeqRecord(seq3)
loop = _get_dna_from_SeqRecord(loop)
pun_xlabel = [i if i % 10 == 0 else '' for i in range(seq_len+len(seq5)+len(seq3)+len(all_uids[0]))]
porp_reduce = (1-(percent_reduce_prob/100))
# find structural regions
regions = [len(seq5), len(seq5)+seq_len,
len(seq5)+seq_len+num5hang+num5polyA+(2*num_bp)+len(loop)]
regionA = list(range(regions[0], regions[1]))