Toggle navigation
Toggle navigation
This project
Loading...
Sign in
Carlos-Francisco Méndez-Cruz
/
deep-learning-workshop
Go to a project
Toggle navigation
Toggle navigation pinning
Projects
Groups
Snippets
Help
Project
Activity
Repository
Pipelines
Graphs
Issues
0
Merge Requests
0
Wiki
Snippets
Network
Create a new issue
Builds
Commits
Issue Boards
Authored by
Carlos-Francisco Méndez-Cruz
2019-05-08 13:45:03 -0500
Browse Files
Options
Browse Files
Download
Email Patches
Plain Diff
Commit
ec0535991209cafb917103827b84f6aa90e50692
ec053599
1 parent
cc41ff4f
Deep Learning Workshop
Show whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
6 additions
and
5 deletions
data-sets/get-hga-training-test-py27.py
data-sets/get-hga-training-test-py27.py
View file @
ec05359
...
...
@@ -96,11 +96,12 @@ if __name__ == "__main__":
# One-hot-encoding of sequences
for
sequence
,
label
in
zip
(
sequences
,
labels
):
if
len
(
sequence
)
<
max_length
:
print
(
"sequence: {}"
.
format
(
sequence
))
print
(
"Le falta: {}"
.
format
(
max_length
-
len
(
sequence
)))
sequence_adjust
=
sequence
.
ljust
(
max_length
-
len
(
sequence
),
'X'
)
print
(
"sequence_adjust: {}"
.
format
(
sequence_adjust
))
# print("sequence: {}".format(sequence))
# print("Le falta: {}".format(max_length - len(sequence)))
sequence_adjust
=
sequence
.
ljust
(
max_length
,
'X'
)
# print("sequence_adjust: {}".format(sequence_adjust))
else
:
sequence_adjust
=
sequence
print
(
"Length sequence_adjust: {}"
.
format
(
len
(
sequence_adjust
)))
integer_encoded
=
integer_encoder
.
fit_transform
(
list
(
sequence_adjust
))
integer_encoded
=
np
.
array
(
integer_encoded
)
.
reshape
(
-
1
,
1
)
...
...
Please
register
or
login
to post a comment