1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
|
# Author: Christoph Helma, Andreas Maunz
module OpenTox
class Application < Service
@warnings = []
helpers do
def from_csv(csv)
table = CSV.parse(csv)
# CSVs with unexpected encodings may have blanks instead of nil
table.collect! { |row|
row.collect! { |val|
(val.class == String and val.strip == "") ? nil : val
}
}
from_table table
end
def from_spreadsheet spreadsheet
extensions = { Roo::Excel => ".xls", Roo::Excelx => ".xlsx", Roo::Openoffice => ".ods" }
input = params[:file][:tempfile].path + ".xls"
csv_file = params[:file][:tempfile].path + ".csv"
File.rename params[:file][:tempfile].path, input # roo needs "correct" extensions
spreadsheet.new(input).to_csv csv_file # roo cannot write to strings
@body = from_csv File.read(csv_file)
@content_type = "text/plain"
end
=begin
def from_sdf(sdf)
#obconversion = OpenBabel::OBConversion.new
#obmol = OpenBabel::OBMol.new
#obconversion.set_in_and_out_formats "sdf", "inchi"
table = []
properties = []
sdf.each_line { |l| properties << l.to_s if l.match(/</) }
properties.sort!
properties.uniq!
properties.collect!{ |p| p.gsub(/<|>/,'').strip.chomp }
properties.insert 0, "InChI"
table[0] = properties
rec = 0
sdf.split(/\$\$\$\$\r*\n/).each do |s|
rec += 1
table << []
begin
# TODO: use compound service
compound = OpenTox::Compound.from_sdf sdf
#obconversion.read_string obmol, s
table.last << obconversion.write_string(obmol).gsub(/\s/,'').chomp
rescue
# TODO: Fix, will lead to follow up errors
table.last << "Could not convert structure at record #{rec}) have been ignored! \n#{s}"
end
obmol.get_data.each { |d| table.last[table.first.index(d.get_attribute)] = d.get_value }
end
from_table table
end
=end
def from_table table
@warnings = []
ntriples = ["<#{@uri}> <#{RDF.type}> <#{RDF::OT.Dataset}>."]
ntriples << ["<#{@uri}> <#{RDF.type}> <#{RDF::OT.OrderedDataset}>."]
ntriples << ["<#{@uri}> <#{RDF::DC.date}> \"#{DateTime.now}\"."]
# features
feature_names = table.shift.collect{|f| f.strip}
@warnings << "Duplicate features in table header." unless feature_names.size == feature_names.uniq.size
compound_format = feature_names.shift.strip
bad_request_error "#{compound_format} is not a supported compound format. Accepted formats: URI, SMILES, InChI." unless compound_format =~ /URI|URL|SMILES|InChI/i
features = []
ignored_feature_indices = []
feature_names.each_with_index do |f,i|
values = table.collect{|row| val=row[i+1]; val.strip! unless val.nil?; val }.uniq.compact
types = values.collect{|v| feature_type(v)}.uniq
metadata = {RDF::DC.title => f}
if values.size == 0 # empty feature
elsif values.size <= 5 # max classes
metadata[RDF.type] = [ RDF::OT.NominalFeature, RDF::OT.StringFeature, RDF::OT.Feature ]
metadata[RDF::OT.acceptValue] = values
end
if types.size == 1 and types[0] == RDF::OT.NumericFeature
metadata[RDF.type] = [] unless metadata[RDF.type]
metadata[RDF.type] << [RDF::OT.NumericFeature, RDF::OT.Feature]
else
metadata[RDF.type] = [ RDF::OT.NominalFeature, RDF::OT.StringFeature, RDF::OT.Feature ] # only nominal type for mixed cases
metadata[RDF::OT.acceptValue] = values
end
feature = OpenTox::Feature.find_or_create metadata
features << feature unless feature.nil?
ntriples << "<#{feature.uri}> <#{RDF.type}> <#{RDF::OT.Feature}>."
ntriples << "<#{feature.uri}> <#{RDF::OLO.index}> #{i} ."
end
# compounds and values
compound_uris = []
r = -1
table.each_with_index do |values,j|
compound = values.shift
compound_uri = nil
begin
case compound_format
when /URI|URL/i
compound_uri = compound
when /SMILES/i
c = OpenTox::Compound.from_smiles(compound)
if c.inchi.empty?
@warnings << "Cannot parse #{compound_format} compound '#{compound}' at position #{j+2}, all entries are ignored."
next
else
compound_uri = c.uri
end
when /InChI/i
c = OpenTox::Compound.from_inchi(compound)
if c.inchi.empty?
@warnings << "Cannot parse #{compound_format} compound '#{compound}' at position #{j+2}, all entries are ignored."
next
else
compound_uri = c.uri
end
else
raise "wrong compound format" #should be checked above
end
rescue
@warnings << "Cannot parse #{compound_format} compound '#{compound}' at position #{j+2}, all entries are ignored." # be careful with double quotes in literals! \C in smiles is an illegal Turtle string
next
end
r += 1
compound_uris << compound_uri
unless values.size == features.size
@warnings << "Number of values at position #{j+2} (#{values.size}) is different than header size (#{features.size}), all entries are ignored."
next
end
ntriples << "<#{compound_uri}> <#{RDF.type}> <#{RDF::OT.Compound}>."
ntriples << "<#{compound_uri}> <#{RDF::OLO.index}> #{r} ."
#data_entry_node = "<#{File.join @uri,"dataentry",j.to_s}>" # too slow or not accepted by 4store
data_entry_node = "_:dataentry"+ r.to_s
ntriples << "<#{@uri}> <#{RDF::OT.dataEntry}> #{data_entry_node} ."
ntriples << "#{data_entry_node} <#{RDF.type}> <#{RDF::OT.DataEntry}> ."
ntriples << "#{data_entry_node} <#{RDF::OLO.index}> #{r} ."
ntriples << "#{data_entry_node} <#{RDF::OT.compound}> <#{compound_uri}> ."
values.each_with_index do |v,i|
if v.blank?
@warnings << "Empty value for compound '#{compound}' (row #{r+2}) and feature '#{feature_names[i]}' (column #{i+2})."
next
else
value_node = data_entry_node+ "_value"+ i.to_s
ntriples << "#{data_entry_node} <#{RDF::OT.values}> #{value_node} ."
ntriples << "#{value_node} <#{RDF::OT.feature}> <#{features[i].uri}> ."
ntriples << "#{value_node} <#{RDF::OT.value}> \"#{v}\" ."
end
end
end
compound_uris.duplicates.each do |uri|
positions = []
compound_uris.each_with_index{|c,i| positions << i+1 if !c.blank? and c == uri}
@warnings << "Duplicate compound #{uri} at rows #{positions.join(', ')}. Entries are accepted, assuming that measurements come from independent experiments."
end
ntriples << "<#{@uri}> <#{RDF::OT.Warnings}> \"#{@warnings.join('\n')}\" ."
ntriples.join("\n")
end
=begin
def to_xlsx
# both simple_xlsx and axlsx create empty documents with OLE2 errors
xlsx = @uri.split("/").last+".xlsx"
p = Axlsx::Package.new
wb = p.workbook
wb.add_worksheet(:name => "test") do |sheet|
to_table.each { |row| sheet.add_row row }
end
p.serialize("test.xlsx")
p.to_stream
#```
#Tempfile.open(@uri.split("/").last+".xlsx") do |xlsx|
SimpleXlsx::Serializer.new(xlsx) do |doc|
doc.add_sheet("People") do |sheet|
to_table.each { |row| sheet.add_row row }
end
end
send_file xlsx
#end
end
=end
def to_csv
csv_string = CSV.generate do |csv|
to_table.each { |row| csv << row }
end
csv_string.gsub(/\"\"/,"") # AM: no quotes for missing values
#to_table
end
def to_table
# TODO: fix and speed up
sparql = "SELECT DISTINCT ?s FROM <#{@uri}> WHERE {
?s <#{RDF.type}> <#{RDF::OT.Feature}> ;
<#{RDF::OLO.index}> ?fidx
} ORDER BY ?fidx"
features = FourStore.query(sparql, "text/uri-list").split("\n").collect { |uri| OpenTox::Feature.new uri }
sparql = "SELECT DISTINCT ?compound FROM <#{@uri}> WHERE {
?compound <#{RDF.type}> <#{RDF::OT.Compound}> ;
<#{RDF::OLO.index}> ?cidx;
} ORDER BY ?cidx"
inchis = FourStore.query(sparql, "text/uri-list").split("\n").collect { |uri| "InChI#{uri.split("InChI").last}" }
table = [["InChI"] + features.collect{|f| f.title}]
inchis.each{|inchi| table << [inchi]}
sparql = "SELECT ?cidx ?fidx ?value FROM <#{@uri}> WHERE {
?data_entry <#{RDF::OLO.index}> ?cidx ;
<#{RDF::OT.values}> ?v .
?v <#{RDF::OT.feature}> ?f;
<#{RDF::OT.value}> ?value .
?f <#{RDF::OLO.index}> ?fidx.
} ORDER BY ?fidx ?cidx"
FourStore.query(sparql,"text/uri-list").split("\n").each do |row|
r,c,v = row.split("\t")
table[r.to_i+1][c.to_i+1] = v.to_s
end
table
end
def feature_type(value)
if value.blank?
nil
elsif value.numeric?
RDF::OT.NumericFeature
else
RDF::OT.NominalFeature
end
end
def ordered?
sparql = "SELECT DISTINCT ?s FROM <#{@uri}> WHERE {?s <#{RDF.type}> <#{RDF::OT.OrderedDataset}>}"
FourStore.query(sparql, "text/uri-list").split("\n").empty? ? false : true
end
def parse_put
task = OpenTox::Task.run "Dataset upload", @uri do
case @content_type
when "text/plain", "text/turtle", "application/rdf+xml" # no conversion needed
when "text/csv", "text/comma-separated-values"
@body = from_csv @body
@content_type = "text/plain"
when "application/vnd.ms-excel", "application/excel"
from_spreadsheet Roo::Excel
when "application/vnd.openxmlformats-officedocument.spreadsheetml.sheet"
from_spreadsheet Roo::Excelx
when "application/vnd.oasis.opendocument.spreadsheet"
from_spreadsheet Roo::Openoffice
# when "chemical/x-mdl-sdfile"
# @body = parse_sdf @body
# @content_type = "text/plain"
else
bad_request_error "#{@content_type} is not a supported content type."
end
if params["file"]
nt = "<#{@uri}> <#{RDF::DC.title}> \"#{params["file"][:filename]}\".\n<#{@uri}> <#{RDF::OT.hasSource}> \"#{params["file"][:filename]}\"."
FourStore.put(@uri, nt, "text/plain")
end
nt ? FourStore.post(@uri, @body, @content_type) : FourStore.put(@uri, @body, @content_type)
@uri
end
response['Content-Type'] = "text/uri-list"
halt 202, task.uri
end
end
end
end
|