repo_name
stringlengths
7
111
__id__
int64
16.6k
19,705B
blob_id
stringlengths
40
40
directory_id
stringlengths
40
40
path
stringlengths
5
151
content_id
stringlengths
40
40
detected_licenses
list
license_type
stringclasses
2 values
repo_url
stringlengths
26
130
snapshot_id
stringlengths
40
40
revision_id
stringlengths
40
40
branch_name
stringlengths
4
42
visit_date
timestamp[ns]
revision_date
timestamp[ns]
committer_date
timestamp[ns]
github_id
int64
14.6k
687M
star_events_count
int64
0
209k
fork_events_count
int64
0
110k
gha_license_id
stringclasses
12 values
gha_fork
bool
2 classes
gha_event_created_at
timestamp[ns]
gha_created_at
timestamp[ns]
gha_updated_at
timestamp[ns]
gha_pushed_at
timestamp[ns]
gha_size
int64
0
10.2M
gha_stargazers_count
int32
0
178k
gha_forks_count
int32
0
88.9k
gha_open_issues_count
int32
0
2.72k
gha_language
stringlengths
1
16
gha_archived
bool
1 class
gha_disabled
bool
1 class
content
stringlengths
10
2.95M
src_encoding
stringclasses
5 values
language
stringclasses
1 value
is_vendor
bool
2 classes
is_generated
bool
2 classes
length_bytes
int64
10
2.95M
extension
stringclasses
19 values
num_repo_files
int64
1
202k
filename
stringlengths
4
112
num_lang_files
int64
1
202k
alphanum_fraction
float64
0.26
0.89
alpha_fraction
float64
0.2
0.89
hex_fraction
float64
0
0.09
num_lines
int32
1
93.6k
avg_line_length
float64
4.57
103
max_line_length
int64
7
931
Rohit-dev-coder/MyPythonCodes
7,834,020,349,528
fe2cf78ce9e1aeea0eff879d0b0cb72927043731
7de2bc9c439bb14f45840dbe9535cb4c65b87320
/upperorlowerletter.py
f7282e4a3993d273f6aa33b9b211962d5ad3086c
[]
no_license
https://github.com/Rohit-dev-coder/MyPythonCodes
51899330ddec120abd3340da4130ac375b74ba25
986fb8d9f40a55d2718563fb2ece3088f49f813b
refs/heads/master
2020-06-29T08:11:21.971783
2019-08-04T11:00:04
2019-08-04T11:00:04
200,481,704
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
ch = input("Enter a Character : ") ch = ch[0] if "A" <= ch <="Z": print("Character is Upper") elif "0" <= ch <= "9": print("Character is Number") else: print("Character is lower")
UTF-8
Python
false
false
189
py
43
upperorlowerletter.py
42
0.57672
0.560847
0
8
22.625
34
carriercomm/Trieste
17,841,294,180,890
9bd60d09008239bb428329e8e5ea94f3e86c2e35
ad4decd93a1baaf700eaee3f4a62085e30765c84
/Trieste/vice/vicedata.py
65b56a281768be8a73f4a7be4769de9635c5d60b
[]
no_license
https://github.com/carriercomm/Trieste
ca666a3cee249f247cc26dcb39d2ce738a394e2f
5891cc1c02f4df9b460c00965c0c5f93f7734d3f
refs/heads/master
2018-02-06T15:36:22.411758
2016-02-24T18:10:04
2016-02-24T18:10:04
null
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
########################################################################### # Copyright (C) 2003 by Marcos Dione # <mdione@grulic.org.ar> # # Copyright: See COPYING file that comes with this distribution # ########################################################################### from bsddb import hashopen from threading import RLock from errno import * from os.path import dirname from os import stat, makedirs, unlink from Trieste.common.utils import csvPrettyPrint, csvParse from Trieste.common.monitor import Monitor, CLock from Trieste.common.object import Object from Trieste.common import consts class ViceData (Object): """ holds three hashes: one for data, anoother for stat, the last for locks the first two are really bsddb's the data one contains three kinds on entries: * entries with key in the from "ino/chunkNo", which are for regular files. each file is broken in chunks of the same size. * entries with key in the from "ino/fileName", which are for directories. that entry means that the dir w/ inode number ino has a link to another ino with filename fileName. the value of such entry is the file's ino. * entries with key in the from "ino", which are also for directories. these ones contain a csv string of the filenames. this one is needed because we need to get the listing of the dir, and it's easier than to iterate over the keys of an bssdb. """ def __init__ (self, master, path): Object.__init__ (self) self.master= master dir= dirname (path) try: stat (dir) except OSError: makedirs (dir) self.path= path self.data= None self.meta= None self.locks= None self.mkDB () # master lock. global actions (like sync and emp) must acquire this one. self.masterLock= RLock () def mkDB (self): path= self.path self.debug (1, "opening %s-data.bsddb" % path) self.data= Monitor (hashopen ("%s-data.bsddb" % path, 'c'), 'data') self.debug (1, "opening %s-metadata.bsddb" % path) self.meta= Monitor (hashopen ("%s-metadata.bsddb" % path, 'c'), 'meta') # per inode locks. any action over an inode must acquire self.locks= Monitor ({}, 'locks') def rmDB (self): self.data= None self.meta= None self.locks= None path= self.path unlink ("%s-data.bsddb" % path) unlink ("%s-metadata.bsddb" % path) def hasInode (self, ino): # self.debug (1, 'looking for inode %s' % ino) self.meta.acquire () ans= self.meta.has_key (str(ino)) self.meta.release () return ans def stat (self, ino, stat=None): if stat: # self.debug (2, 'setting stat') self.meta[str(ino)]= str(stat) else: stat= eval (self.meta[str(ino)]) # TODO: so bark if it's not right return stat def keys (self): self.meta.acquire () l= map (lambda x: int(x), self.meta.keys ()) self.meta.release () return l def getValues (self, key): # needed for master.giveData() # this way a vice gives his url as the data (to be added [to the list that it's there]) # but it's a very ugly generalization return [self.master.url ()] ######### # file ops ######### def size (self, ino, size=None): stat= self.stat (ino) if size is not None: stat[consts.statSize]= size self.stat (ino, stat) return stat[consts.statSize] def pages (self, ino): """ How much pages does ino have allocated. """ # for size==0 return 0 # for sizes E (1..pageSize) return 1 # and so forth size= self.size (ino) pages= (size+consts.pageSize-1)/consts.pageSize # self.debug (2, "pages: %d, size: %d" % (pages, size)) return pages def read (self, ino, off=0, size=None): # profiling! # return "\0"*size # lot of things: # first, if data is spread in several pages, read them # second, what if @ eof? None if size is None: size= self.size (ino) # self.debug (2, "reading @ %d, size %d" % (off, size)) data= '' bytesToRead= size # ugh bytesRead= size try: while bytesToRead and bytesRead: (page, start)= divmod (off, consts.pageSize) end= min (consts.pageSize, start+bytesToRead) pageData= self.data["%d/%d" % (ino, page)] readData= pageData[start:end] bytesRead= len (readData) # self.debug (2, "off: %d, chunk: %d, chunkOff(CO): %d, max: %d, bytesRead: %d, bytesTR: %d, pageSz-CO: %d, CO+bytesTR: %d" % # # (off, page, start, end, bytesRead, bytesToRead, consts.pageSize-start, start+bytesToRead)) data+= readData bytesToRead-= bytesRead off+= bytesRead # self.debug (2, "off: %d, chunk: %d, chunkOff(CO): %d, max: %d, bytesRead: %d, bytesTR: %d, pageSz-CO: %d, CO+bytesTR: %d\n" % # (off, page, start, end, bytesRead, bytesToRead, consts.pageSize-start, start+bytesToRead)) except KeyError: # means EOF? pass return data def trunc (self, ino, size=0, shrink=True): # self.debug (2, 'trunc\'ing', 2) fileSize= self.size (ino) if size==fileSize: # self.debug (2, 'already there (%d<->%d); bailing out' % (size, fileSize)) return size # check if we need to create previous pages lastPage= self.pages (ino) (page, end)= divmod (size, consts.pageSize) if page>lastPage: # self.debug (1, 'expanding') # ... data | # ... | lastPage | ... | page | # ^ end # first add '0's to last page try: pageData= self.data["%d/%d" % (ino, lastPage)] except KeyError: self.master.fblocks-= 1 if self.master.fblocks==0: return -ENOSPC pageData= '' # decrease free count pageLen= len (pageData) if pageLen<consts.pageSize: # self.debug (1, 'filling up lastPage: %d-> %d' % (pageLen, consts.pageSize)) pageData+= "\0"*(consts.pageSize-len (pageData)) self.data["%d/%d" % (ino, lastPage)]= pageData # ... data |0000| # ... | lastPage | ... | page | # ^ end # now fill the pages gap, starting from the next one to the last # till the previus one to the wanted page i= lastPage+1 while i<page: self.master.fblocks-= 1 if self.master.fblocks==0: return -ENOSPC # self.debug (1, 'added page %d' % i) self.data["%d/%d" % (ino, i)]= "\0"*consts.pageSize # decrease free count i+= 1 # ... data |000000...0| # ... | lastPage | ... | page | # ^ end # decrease free count self.master.fblocks-= 1 # self.debug (1, 'filling up page: -> %d' % end) self.data["%d/%d" % (ino, page)]= "\0"*end elif page<=lastPage: # ... data | # ... | page | ... | lastPage | # ^ end if shrink: # self.debug (1, 'shrinking') i= lastPage while i>page: try: del self.data["%d/%d" % (ino, i)] # self.debug (1, 'del\'ed page %d' % i) except KeyError: # self.debug (1, 'page %d not present for deletion' % i) pass # increase free count self.master.fblocks+= 1 i-= 1 # self.debug (1, 'done sh\'k\'n') try: pageData= self.data["%d/%d" % (ino, page)] except KeyError: pageData= '' # decrease free count self.master.fblocks-= 1 pageLen= len(pageData) if pageLen>end and shrink: # ... data | # ... | page | ... | lastPage | # ^ end pageData= pageData[:end] else: # ... data | # ... | page | ... | lastPage | # ^ end pageData+= "\0"*(end-pageLen) # self.debug (1, 'somehting\'ing page: %d-> %d' % (pageLen, end)) self.data["%d/%d" % (ino, page)]= pageData # ... data | # ... | page | # ^ end # modify size if (shrink and size<fileSize) or size>fileSize: # self.debug (1, 'change size: %d-> %d' % (fileSize, size)) self.size (ino, size) return size def write (self, ino, off, data): # profiling # return len(data) # self.debug (2, "writing in %d @ %d, size %d" % (ino, off, len(data))) bytesToWrite= len (data) totalBytesWrote= 0 self.trunc (ino, off, False) while bytesToWrite: # start end # | | # v v # ... | page | ... (page, start)= divmod (off, consts.pageSize) end= min (consts.pageSize, start+bytesToWrite) # self.debug (2, "o %d; btw %d; p %d[%d..%d]" % (off, bytesToWrite, page, start, end)) # get the page we'll be writing to try: pageData= self.data["%d/%d" % (ino, page)] except KeyError: # decrease free count self.master.fblocks-= 1 if self.master.fblocks==0: return -ENOSPC # self.debug (2, 'new page %d' % page) pageData= '' pageLen= len(pageData) # write bytesWrote= end-start # self.debug (2, ">%s<" % pageData) # self.debug (2, "page: %d, start: %d, bytesToWrite: %d, bytesWrote: %d, end: %d, page[->]: >%s<, data[-]: >%s<, page[<-]: >%s<" % # (page, start, bytesToWrite, bytesWrote, end, pageData[:start], data[:bytesWrote], pageData[start+bytesWrote:pageLen])) pageData= pageData[:start]+data[:bytesWrote]+pageData[start+bytesWrote:pageLen] # self.debug (2, ">%s<" % pageData) self.data["%d/%d" % (ino, page)]= pageData # adjust indexes and remaining data data= data[bytesWrote:] bytesToWrite-= bytesWrote totalBytesWrote+= bytesWrote off+= bytesWrote # update ino size fileSize= self.size (ino) if off>fileSize: fileSize= self.size (ino, off) # for equalization in tests w/ adversaries # self.data.sync () # self.meta.sync () return totalBytesWrote def lock (self, ino, creating=False): error= -ENOENT # self.locks.acquire () # if not self.locks.has_key (ino): # create lock # self.locks[ino]= CLock () # get lock on ino # self.locks[ino].acquire () exists= self.hasInode (ino) if (exists and not creating) or (not exists and creating): error= 0 else: if exists and creating: error= -EEXIST # else -ENOENT # self.unlock (ino) # self.locks.release () return error def unlock (self, ino): error= -ENOENT if self.hasInode (ino): error= 0 # self.locks.acquire () # count= self.locks[ino].release () # if not count: # del self.locks[ino] # self.locks.release () return error def sync (self): self.masterLock.acquire () self.debug (1, 'sync') self.data.acquire () self.data.sync () self.data.release () self.meta.acquire () self.meta.sync () self.meta.release () self.masterLock.release () ############# # dir methods ############# def mkdir (self, ino, parent): # decrease free counts self.master.fblocks-= 1 self.dirContents (ino, []) # completeness self.link (ino, '.', ino, True, False) # link already adds the name to the list above self.link (ino, '..', parent, True, False) def rmdir (self, ino): error= -ENOTEMPTY # delete both ino and all ino/child... in reverse order, obv. children= self.dirContents (ino) if len (children)==2: for child in children: key= "%d/%s" % (ino, child) del self.data[key] # now ino... del self.data["%d" % ino] # and metadata del self.meta["%d" % ino] # dec file count self.master.ufiles-= 1 # inc free counts self.master.fblocks+= 1 def link (self, dirIno, fileName, fileIno, inc, over): self.debug (1, 'linking %d:%s:%d' % (dirIno, fileName, fileIno)) error= -EEXIST key= "%d/%s" % (dirIno, fileName) if not self.data.has_key (key) or over: # add it if it's no there list= self.dirContents (dirIno) list.append (fileName) self.dirContents (dirIno, list) self.data[key]= str(fileIno) if inc: # inc link count stat= self.stat (dirIno) stat[consts.statNlink]+= 1 stat= self.stat (dirIno, stat) error= 0 return error def unlink (self, dirIno, fileName, dec): error= -ENOENT key= "%d/%s" % (dirIno, fileName) if self.data.has_key (key): # bye bye del self.data[key] # and remove it from the list list= self.dirContents (dirIno) # no checks; the try catches it try: list.remove (fileName) self.dirContents (dirIno, list) if dec: # dec link count stat= self.stat (dirIno) stat[consts.statNlink]-= 1 stat= self.stat (dirIno, stat) error= 0 except: # ENOENT pass return error def lookup (self, dirIno, fileName): """ if there's an entry w/ that name, return the inode """ key= "%d/%s" % (dirIno, fileName) if self.data.has_key (key): ans= int(self.data[key]) else: ans= None return ans def dirContents (self, ino, contents=None): if not contents==None: self.data["%d" % ino]= csvPrettyPrint (contents) ans= contents else: ans= csvParse (self.data["%d" % ino]) return ans ####### # misc ####### def emp (self): self.masterLock.acquire () self.data.close () self.data.release () self.meta.close () self.meta.release () self.rmDB () self.mkDB () self.data.acquire () self.meta.acquire () self.masterLock.release () def fragments (self, ino): """ How much pages does ino have allocated. """ # for size==0 return 0 # for sizes E (1..pageSize) return 1 # and so forth size= self.size (ino) fragments= (size+consts.fragmentSize-1)/consts.fragmentSize self.debug (2, "fragments: %d, size: %d" % (fragments, size)) return fragments def usedBlocksAndFiles (self): files= 0 blocks= 0 self.meta.acquire () # sometimes I just hate exceptions... try: key= self.meta.first ()[0] except KeyError: key= None while key and not self.master._terminate: self.debug (1, 'found key %s' % key) files+= 1 blocks+= self.fragments (key) try: key= key= self.meta.next ()[0] except KeyError: key= None self.meta.release () return (blocks, files)
UTF-8
Python
false
false
14,734
py
37
vicedata.py
34
0.558708
0.552464
0
495
28.765657
143
niezongxia/personal
18,090,402,275,959
7b9bbbd859ee869cf138338ec432341ac235b55f
01cfe7cd0f1d24e1224f52b15cf86f223b52692d
/python/code/wait/read_excel.py
ef28a7202b54905ed44c7de1352a03f93be374f2
[]
no_license
https://github.com/niezongxia/personal
a782397068bc2be6cae5e4cac0a0ee9ff4927443
ff70eb1dba1cc8ac47b863e55533902bec1604e1
refs/heads/master
2022-10-02T14:36:20.163143
2022-08-27T14:09:25
2022-08-27T14:09:25
143,152,875
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
#read_excel.py import xlrd fname="C:/Users/NZX/Desktop/test.xls" rbook=xlrd.open_workbook(fname) #row_value=xls_sheet.row_values(num-1) #cow_value=xls_sheet.col_values(num-1) rows=rbook.sheets()[0].nrows#获取已有行数 cols=rbook.sheets()[0].ncols#获取已有列数 for row in rows: value=rbook.cell_value(row,1) print(value)
UTF-8
Python
false
false
347
py
25
read_excel.py
25
0.718266
0.702786
0
19
16
38
KouheiFurukawa/atcoder-back-number
5,093,831,240,944
25b13fe4b92ae9c83a8f227eb508164e2e4a8a39
b27b0bcd22973d029688ea3f10c9abfe651800a4
/abc137/abc_137_c.py
8c35b7c06b10186c08102d4a0a323e50017c0777
[]
no_license
https://github.com/KouheiFurukawa/atcoder-back-number
8f6b7c6704bc6fb32bb8ca9ccc9971650c0fdf8f
2fe11830614bf2ce64ee7947734043fe47800277
refs/heads/master
2020-06-03T13:45:54.531015
2020-05-17T05:53:39
2020-05-17T05:53:39
191,591,257
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
from sys import stdin N = int(stdin.readline().rstrip()) S = [stdin.readline().rstrip() for _ in range(N)] count = {} for s in S: s = ''.join(sorted(s)) if s not in count: count[s] = 1 else: count[s] += 1 ans = 0 for c in count: if count[c] > 1: ans += count[c] * (count[c] - 1) // 2 print(ans)
UTF-8
Python
false
false
339
py
237
abc_137_c.py
235
0.519174
0.501475
0
19
16.842105
49
olivercalder/scherzi
12,034,498,374,276
45567a122cc837e49dc26ee2bb74300de368f8fe
32a8b94476ab9fcb4c3896ebd1ac050d5389ab52
/44.py
27af49538181cd34a190bc5249829326c23e3670
[]
no_license
https://github.com/olivercalder/scherzi
abf408c27a48dc3122b451768de76b937f2f0886
5a1fb78bf024d5176c0e100a818cc7205b135281
refs/heads/master
2022-05-02T17:12:26.138819
2022-04-26T04:00:26
2022-04-26T04:00:26
212,889,121
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
def add_link(path, n): if len(path) == 2*n: return [path] new_paths = [] if path.count('u') < n: u_path = path.copy() u_path.append('u') new_paths += add_link(u_path, n) if path.count('r') < n: r_path = path.copy() r_path.append('r') new_paths += add_link(r_path, n) return new_paths def count_above(path): total = 0 for i in range(len(path)): link = path[i] if link == 'u' and path[:i].count('u') >= (i+1)//2: total += 1 return total def flip_path(path, u_num): i = 0 count = 0 while count < u_num: if path[i] == 'u': count += 1 i += 1 path = [] n3_paths = add_link(path, 3) print(len(n3_paths)) for p in n3_paths: print(''.join(p)) print(count_above(p))
UTF-8
Python
false
false
823
py
49
44.py
41
0.487242
0.471446
0
36
21.861111
59
adderan/VBS
5,050,881,548,698
0a79c2f38e8dea61f0ffeb0b962ce57318e71aaa
a598a793998408416107615717eafa96c4efcd18
/DelphesResolution/scripts/res_formula.py
22d0df13ce7ec359f64fb67fee340d2ffc8aa67d
[]
no_license
https://github.com/adderan/VBS
f1bce5ed36eb65fdcbe5778746752ef6b3cb6157
85ae92d48797f4f3d91e14345f788a9cfb200cbc
refs/heads/master
2016-02-24T17:54:51.557152
2015-02-24T06:15:55
2015-02-24T06:15:55
26,232,911
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
import math, formula_reader def jet_res(PT, Eta, pileup = 0): parameters = args.parameters.readlines() a = None b = None S = None C = None #print "Eta = " + str(Eta) for line in parameters: #print(line) pars = line.split() etaregion = float(pars[0]) a = float(pars[1]) b = float(pars[2]) S = float(pars[3]) C = float(pars[4]) if Eta < etaregion: break N = a + b*pileup sigma = math.sqrt((N**2)/(PT**2) + (S**2)/PT + C**2) return sigma def sigma_id(PT, eta): parameters = formula_reader.read_parameters("parameters/idparameters.csv") eta_step = float(parameters[0][1]) - float(parameters[0][0]) parameter_index = int(eta/eta_step) eta1, eta2, a1, a2 = parameters[parameter_index] return math.sqrt(float(a1)*float(a1) + float(a2)*PT*float(a2)*PT) def sigma_ms(PT, eta): parameters = [[0.24, 0.02676, 0.00012], [0.0, 0.0388, 0.00016]] if eta < 1.05: parameter_index = 0 else: parameter_index = 1 b0, b1, b2 = parameters[parameter_index] return math.sqrt((float(b0)/PT)**2 + float(b1)**2 + (float(b2)*PT)**2) def sigma_linear(PT, eta): parameters = 0.2 return parameters def sigma_idms(PT, eta): id_value = sigma_id(PT, eta) ms_value = sigma_ms(PT, eta) #print "MS = " + str(ms_value) + " ID = " + str(id_value) return id_value * PT * ms_value / math.sqrt((PT*id_value)**2 + (PT*ms_value)**2)
UTF-8
Python
false
false
1,344
py
9
res_formula.py
5
0.634673
0.584077
0
45
28.822222
81
iamtrueline/Programmers
438,086,678,681
e9f725da941f5c2a3320fbfd863efd31cf019697
9eacf21048e9224443d7a7775497adf824f5549b
/Python/조이스틱.py
53c45f09c6a521e09c24a0238b6e693fb5ba4b1d
[]
no_license
https://github.com/iamtrueline/Programmers
fb0eb29046ba813aa532a35b474ae8349c827e62
ac19a3999cc46c748593ab9e08f4d14a13579312
refs/heads/main
2023-06-26T05:34:44.899022
2021-07-30T03:23:02
2021-07-30T03:23:02
380,900,305
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
def solution(name): answer = 0 change = [min(ord(i)-ord('A'), ord('Z')-ord(i)+1) for i in name] idx = 0 while True: answer += change[idx] change[idx]=0 left = 1 right = 1 if sum(change)== 0: return answer while change[idx- left] == 0: left += 1 while change[idx + right] == 0: right += 1 answer += min(left, right) idx += -left if left < right else right return answer
UTF-8
Python
false
false
526
py
40
조이스틱.py
39
0.452471
0.431559
0
23
21.869565
68
luoxz/GID_Internal
730,144,482,669
d864ae640e6a82699c7635c16bfc6ef0dc5b98f9
0010f2f71aa04595ecbc84691f330c185f5c7356
/test/client/test_dev_server.py
87f883bd62c84d3f91aa028690e9f8f8b9f8a636
[ "MIT" ]
permissive
https://github.com/luoxz/GID_Internal
e38be906f255931e6ef2a1a03695ee33db3f804b
452cca7543d80f31c8f659709d9143ecc1d8eb93
refs/heads/master
2021-01-16T00:56:26.411020
2017-07-18T18:55:01
2017-07-18T18:55:01
99,984,036
1
0
null
true
2017-08-11T02:37:40
2017-08-11T02:37:40
2017-07-12T20:35:47
2017-07-18T18:57:01
35,094
0
0
0
null
null
null
''' Test the correctness of the dev server, which simulates the functions of unrealcv server in python ''' import threading, time, socket, unittest, logging, sys if (sys.version_info > (3, 0)): import socketserver as SocketServer else: import SocketServer from dev_server import MessageServer, MessageTCPHandler import unrealcv host = 'localhost' port = 9001 def test_server(): server = MessageServer((host, port)) server.start() server.shutdown() def test_release(): ''' Test whether resources are correctly released ''' for i in range(10): server = MessageServer((host, port)) server.start() # Make sure the port has been released, otherwith this will fail s = socket.socket(socket.AF_INET, socket.SOCK_STREAM) s.connect((host, port)) server.shutdown() def test_client_side_close(): ''' Test whether the server can correctly detect client disconnection ''' server = MessageServer((host, port)) server.start() for i in range(10): s = socket.socket(socket.AF_INET, socket.SOCK_STREAM) s.connect((host, port)) unrealcv.SocketMessage.WrapAndSendPayload(s, 'hello') s.close() # It will take some time to notify the server time.sleep(0.5) # How long will the server should detect the client side loss server.shutdown() if __name__ == '__main__': test_server() test_release() test_client_side_close()
UTF-8
Python
false
false
1,460
py
42
test_dev_server.py
22
0.664384
0.656164
0
51
27.627451
98
cowcai1988/-Poi-
2,851,858,305,504
38cdd581f1d5dd18c739e4d8b6a4324bf79a08e1
50e03c155e196a2d645b7b0d1849e80e495d0a36
/yongdushuju.py
c4817d8537508cef81f3926491a93310b9607677
[]
no_license
https://github.com/cowcai1988/-Poi-
8f36938880ecfbe3c82f30febeb6bf05caaeadb1
764b08961a6343b1760a76c6d22da3880fd581d1
refs/heads/master
2020-07-26T06:01:26.036224
2019-05-20T07:34:37
2019-05-20T07:34:37
null
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
import requests import json import time import pymysql import threading from pandas import DataFrame def getGaodeTrafficStatus(key,furl,currentTime): insert_list = [] TrafficStatusUrl = furl res = requests.get(url=TrafficStatusUrl).content res=res.decode("utf-8") total_json = json.loads(res) print(total_json) jsondata = total_json['trafficinfo']['roads'] currentDate = time.strftime("%Y-%m-%d", time.localtime()) if any(jsondata): for i in jsondata: name = i['name'] status = i['status'] direction = i['direction'] evaluation=total_json['trafficinfo']['evaluation'] df = DataFrame({ 'p_str': total_json['trafficinfo']['evaluation'] }); p_float = df['p_str'].str.strip("%") evaluation=p_float angle = i['angle'] speed = i.get('speed') if speed is None: speed = None lcodes = i['lcodes'] polyline = i['polyline'] list = [name, evaluation,status,direction, angle, lcodes, polyline, currentDate, currentTime, speed] insert_list.append(list) db = pymysql.connect("localhost", "root", "root", "ttt") cursor = db.cursor() print(len(insert_list)) for i in insert_list: print(len(i)) if len(i): print( "insert into biao(name,evaluation,status,direction,angle,lcodes,polyline, currentDate, currentTime, speed) values('%s','%s','%s','%s','%s','%s','%s','%s','%s','%s')" % ( i[0], i[1], i[2], i[3], i[4], i[5], i[6], i[7], i[8], i[9])) print("----------------------------") cursor.execute( "insert into biao(name,evaluation,status,direction,angle,lcodes,polyline, currentDate, currentTime, speed) values('%s','%s','%s','%s','%s','%s','%s','%s','%s','%s')" % ( i[0], i[1], i[2], i[3], i[4], i[5], i[6], i[7], i[8], i[9])) db.commit() db.close() keyList=[{}] rectangleList=[] def pydata(): currentTime = time.strftime("%H:%M:%S", time.localtime()) key='bc411bc209882674362f05b90d6167d1' rectangle='112.89173,23.17571;112.90443,23.16561'#矩形左上右下坐标;隔开两个坐标 # location='112.86461,23.51614'#圆形搜索 # location_arr="[112.86461,23.51614],[112.87079,23.46136],[112.84058,23.41411],[112.94632,23.41159]" # radius='5000'#圆形半径 type='rect' if(type=="rect"): url='http://restapi.amap.com/v3/traffic/status/rectangle?key='+key+'&rectangle='+rectangle+'&extensions=all' else: url='http://restapi.amap.com/v3/traffic/status/circle?key='+key+'&location='+location+'&radius='+radius+'&extensions=all' getGaodeTrafficStatus(key,url,currentTime); timer = threading.Timer(5,pydata) timer.start() if __name__ == "__main__": timer = threading.Timer(5,pydata) timer.start()
UTF-8
Python
false
false
3,165
py
2
yongdushuju.py
1
0.530619
0.479962
0
72
41.319444
189
jiaweiM/PythonNotes
18,433,999,657,422
337516b42f7349c093905f2bff8b8e7aa936b09a
ae6e9d7ab3795d7580ba2d8dfa796e28d94e315c
/src/ml_test/lr_1.py
388019e9e0b5e24abbde492f5e1b96621fbeaeb4
[]
no_license
https://github.com/jiaweiM/PythonNotes
302a3c6d04638628a94a5ccd96b848258c23653c
66227b740cd05d61d77f1cd57a5f4e39f7c1a26f
refs/heads/master
2021-09-22T08:47:42.956762
2021-09-14T07:44:19
2021-09-14T07:44:19
246,194,973
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
import numpy as np import pandas as pd import tensorflow as tf california_housing_dataframe = pd.read_csv("../datasets/california_housing_train.csv", sep=",") california_housing_dataframe = california_housing_dataframe.reindex( np.random.permutation(california_housing_dataframe.index)) california_housing_dataframe['median_house_value'] /= 1000.0 print(california_housing_dataframe.describe()) my_feature = california_housing_dataframe[["total_rooms"]] feature_columns = [tf.feature_column.numeric_column("total_rooms")] targets = california_housing_dataframe["median_house_value"] # Use gradient descent as the optimizer for training the model. my_optimizer = tf.train.GradientDescentOptimizer(learning_rate=0.0000001) my_optimizer = tf.contrib.estimator.clip_gradients_by_norm(my_optimizer, 5.0) # Configure the linear regression model with our feature columns and optimizer. # Set a learning rate of 0.0000001 for Gradient Descent. linear_regressor = tf.estimator.LinearRegressor( feature_columns=feature_columns, optimizer=my_optimizer )
UTF-8
Python
false
false
1,061
py
393
lr_1.py
177
0.787936
0.766258
0
26
39.807692
95
shahrulkamil98/183DB-Capstone
9,646,496,590,658
c0b38865460ad5ef2ab7ab882e12598308b95667
fca8c4124dfbec8975385df924b010829be38a2e
/Demo/Demo/controllers/computer/computer.py
94b9bd0f7e1303090aaecbdb95b9a35bd49399e8
[]
no_license
https://github.com/shahrulkamil98/183DB-Capstone
37a2b56dfbec50488ba44b7f0e9d2dc666700930
4e8dab3e00e55dc8e2dc9242e7c52a4220176b48
refs/heads/main
2023-06-01T16:18:50.144327
2021-06-12T01:52:30
2021-06-12T01:52:30
null
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
# computer.py from controller import * receiverRobot = Robot() TIME_STEP = int(receiverRobot.getBasicTimeStep()) computer_receiver = receiverRobot.getDevice('computer_receiver') computer_emitter = receiverRobot.getDevice('computer_emitter') computer_receiver.enable(TIME_STEP) t = 0 while(t<1000): receiverRobot.step(TIME_STEP) if computer_receiver.getQueueLength() > 0: A = computer_receiver.getData() A_decoded = int.from_bytes(A, byteorder='big') computer_receiver.nextPacket() t = t+TIME_STEP # Next function receiverRobot.step(TIME_STEP) computer_emitter.send(bytes([2])) computer_emitter.send(bytes([1])) computer_emitter.send(bytes([4])) computer_emitter.send(bytes([3])) while t<10000: receiverRobot.step(TIME_STEP)
UTF-8
Python
false
false
741
py
20
computer.py
7
0.755735
0.735493
0
31
22.935484
64
Rao-Varun/gamerbuddy
16,630,113,372,195
cedb0416e5b1158c0ff2d6eed89c619e9972486e
2dd7a5964e6333217fb816e8164f120d612c6098
/venv/Lib/site-packages/nltk/test/unit/test_wordnet.py
f83dbcca93d26ed64124b4e9ce8e431b8a1f0b6a
[]
no_license
https://github.com/Rao-Varun/gamerbuddy
ab8cf2ab19f113edc1a3b0626c8645a30b910fcd
01517ead2685789a4de73a5d696c33c30b2fc6d2
refs/heads/master
2020-05-02T00:48:34.195704
2019-05-09T00:00:48
2019-05-09T00:00:48
177,677,039
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
version https://git-lfs.github.com/spec/v1 oid sha256:09c70093103df502db9a37d82de634943533749942d07e818f46a81529abe41f size 7066
UTF-8
Python
false
false
129
py
250
test_wordnet.py
228
0.883721
0.457364
0
3
42
75
redbrick/rbvm-libvirt
15,238,544,002,141
3b322126145960ab7db87eb15a4c5f2bc7b0df77
462b8d696e9e05322ba0b48568768bf1b5850513
/rbvm/model/database.py
9f7b31c192687b429965b33908fba44a94616642
[]
no_license
https://github.com/redbrick/rbvm-libvirt
fbedd656d8a3fbcf84fb20e7ae7286b042a8b69a
bd6f75964fb2f01cb5bb3d6facfc0857e4c3d11c
refs/heads/master
2021-01-22T06:18:38.589221
2017-05-26T19:56:51
2017-05-26T19:56:51
92,536,318
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
# coding=utf-8 import sqlalchemy import random import string import hashlib import os import base64 import datetime import libvirt from sqlalchemy import Table,Column,MetaData,ForeignKey from sqlalchemy.schema import Sequence, ForeignKeyConstraint from sqlalchemy import Integer,String,DateTime,Unicode,SmallInteger,Text,Binary,Boolean from sqlalchemy.ext.declarative import declarative_base from sqlalchemy.orm import relation,backref from sqlalchemy.ext.associationproxy import association_proxy import rbvm.lib.sqlalchemy_tool as database session = None # Initialised at runtime by single-threaded daemons (multi threaded daemons use sqlalchemy_tool) Base = declarative_base() class User(Base): """ User """ # {{{ __tablename__ = 'user_table' id = Column(Integer,Sequence('user_table_id_seq'),primary_key=True) username = Column(String(255),unique=True,nullable=False) salt = Column(String(10),nullable=False) password = Column(String(255),nullable=False) email_address = Column(String(255),nullable=False) domains = relation('Domain', backref='user') def set_password(self,password_plain): salt = ''.join(random.Random().sample(string.letters + string.digits,9)) hash = hashlib.sha256() hash.update(password_plain + salt) self.password = hash.hexdigest() self.salt = salt def __repr__(self): return "<User('%s')>" % (self.username) def __init__(self,username,email_address,password_plain=None): self.username = username if not password_plain: password_plain = "".join(random.sample(string.letters + string.digits,8)) self.set_password(password_plain) self.email_address = email_address def has_ability(ability_name): for ability_obj in self.abilities: if ability_obj.system_name == ability_name: return True for group in self.groups: if group.has_ability(ability_name): return True return False # }}} user_group = Table('user_group',Base.metadata, # {{{ Column('user_id',Integer,ForeignKey('user_table.id')), Column('group_id',Integer,ForeignKey('group_table.id')) ) # }}} class Group(Base): """ User gorup """ # {{{ __tablename__ = 'group_table' id = Column(Integer,Sequence('group_table_id_seq'),primary_key=True) name = Column(String(255)) system_name = Column(String(255)) users = relation('User',secondary=user_group,backref='groups') def __repr__(self): return "<Group('%s')>" % (self.name) def __init__(self,name,system_name): self.name = name self.system_name = system_name def has_ability(self, ability_name): for ability_obj in self.abilities: if ability_obj.system_name == ability_name: return True return False # }}} user_ability = Table('user_ability',Base.metadata, # {{{ Column('user_id',Integer,ForeignKey('user_table.id')), Column('ability_id',Integer,ForeignKey('ability.id')) ) # }}} group_ability = Table('group_ability',Base.metadata, # {{{ Column('group_id',Integer,ForeignKey('group_table.id')), Column('ability_id',Integer,ForeignKey('ability.id')) ) # }}} class Ability(Base): """ Abilities, assigned to groups and users """ # {{{ __tablename__ = 'ability' id = Column(Integer,Sequence('ability_id_seq'),primary_key=True) name = Column(String(255)) system_name = Column(String(255)) users = relation('User',secondary=user_ability,backref="abilities") groups = relation('Group',secondary=group_ability,backref="abilities") def __repr__(self): return "<Ability('%s')>" % self.system_name def __init__(self, name, system_name): self.name = name self.system_name = system_name # }}} class Hypervisor(Base): """ A hypervisor """ __tablename__ = 'hypervisor' id = Column(Integer,Sequence('hypervisor_id_seq'),primary_key=True) name = Column(String(255)) uri = Column(String(1024)) domains = relation('Domain',backref='hypervisor') def __repr__(self): return "<Hypervisor('%s')>" % self.uri def __init__(self, name, uri): self.name = name self.uri = uri def connect(self): return libvirt.open(self.uri) def list_domains(self): conn = self.connect() return [conn.lookupByName(n) for n in conn.listDefinedDomains()] + [conn.lookupByID(i) for i in conn.listDomainsID()] class Domain(Base): """ Maps a domain (by UUID) to a user. """ __tablename__ = 'domain' id = Column(Integer,Sequence('domain_id_seq'),primary_key=True) uuid = Column(String(36)) user_id = Column(ForeignKey('user_table.id')) hypervisor_id = Column(ForeignKey('hypervisor.id')) def __repr__(self): return "<Domain('%s'>)" % (self.uuid) def __init__(self, uuid, user, hypervisor): self.uuid = uuid self.user_id = user.id self.hypervisor_id = hypervisor.id class OneTimeToken(Base): """ A token that can be sent to the client (in unreadable form) and sent back to verify a command's origin. """ # {{{ __tablename__ = 'one_time_token' id = Column(Integer,Sequence('one_time_token_id_seq'),primary_key=True) token = Column(String(255),index=True) timestamp = Column(DateTime) used = Column(Boolean) user_id = Column(ForeignKey('user_table.id')) def __repr__(self): return "<OneTimeToken('%s')>" % (self.token) def __init__(self,user): assert user is not None # Generate a random token self.token = base64.b64encode(os.urandom(200))[:255] self.timestamp = datetime.datetime.now() self.used = False self.user_id = user.id def check_and_expire(self,user): """ Returns whether or not a token has been used before or is invalid, and marks the token as used. """ seconds = 60 * 15 delta = datetime.timedelta(seconds=seconds) try: assert user is not None assert self.user_id == user.id assert self.used == False and self.timestamp + delta > datetime.datetime.now() except AssertionError: return True self.used = True if session is None: database.session.commit() else: session.commit() return False # }}}
UTF-8
Python
false
false
6,766
py
41
database.py
35
0.602128
0.593704
0
235
27.791489
125
apeckman/PennApps2015
13,950,053,825,335
4c4ddf8b1ac08434123b1c9d3533572335546caf
0548becc751743b9243fe99eb6c68ad4b835649e
/bae/stocks/PythonAlg/Ftl.py
3760d8b1421221bd2355f70124a7e35c71a9fbb8
[]
no_license
https://github.com/apeckman/PennApps2015
0deda7cccc026dffdd699ceef1a61ce76cf90042
0f88b64fea91dbbe0c39348424788c435a580e27
refs/heads/master
2021-01-15T19:28:00.989846
2015-09-06T10:35:37
2015-09-06T10:35:37
41,937,692
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
import sys class Ftl(object): exp = [] algWeight = 1.0 name = 'FTL' def __init__(self): self.exp = [] def getName(self): return self.name def getWeight(self): return self.algWeight def setWeight(self,_weight): self.algWeight = _weight def add(self, e): self.exp.append(e) def ftl(self): name = None maxf = sys.float_info.max for e in self.exp: if (e.cost < maxf): name = e maxf = e.cost return name
UTF-8
Python
false
false
447
py
31
Ftl.py
20
0.61745
0.612975
0
30
13.8
29
PyBeaner/IntroductionToAlgorithms
8,040,178,818,791
e8a4c70ba654c1010aca17e540592f948fcf3d98
896c93a239de381c40151832b4d07b538bc86d63
/第一部分 基础知识/第四章 分治策略/4.1 最大子数组/max-subarray-brute-force.py
a965a79bc5c65d91403dffc613d7666f5041c677
[]
no_license
https://github.com/PyBeaner/IntroductionToAlgorithms
59b26fd8ab25202eaeca0745e5f4be3bbadb89aa
561b0d82540769a1cc787226b44e9c2e44d28925
refs/heads/master
2020-05-18T10:22:32.577205
2015-07-30T15:53:24
2015-07-30T15:53:24
38,962,003
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
__author__ = 'PyBeaner' threshold = -10**10 # O(n^2) def find_max_subarray(alist): length = len(alist) max_sum = threshold left = right = 0 for i in range(length): left_in_loop = right_in_loop = i sum_in_loop = 0 max_sum_in_loop = threshold for j in range(i, length): sum_in_loop += alist[j] if sum_in_loop > max_sum_in_loop: max_sum_in_loop = sum_in_loop right_in_loop = j if max_sum_in_loop > max_sum: max_sum = max_sum_in_loop left = left_in_loop right = right_in_loop return left, right, max_sum if __name__ == '__main__': from random import sample # alist = sample(range(-10, 0), 10) alist = sample(range(-10, 10), 10) print(alist) result = find_max_subarray(alist) print(result)
UTF-8
Python
false
false
866
py
29
max-subarray-brute-force.py
29
0.533487
0.512702
0
33
25.242424
45
gitddabong/acmicpc
1,640,677,544,517
1ebc639459bd062bb69aeec921d5636e0ad2967a
a7ebd5ac27af32faa19ea84b9a2b64abdf70ae22
/2751.py
e79bc9206329f8ed31d0d1944970211b2ff69d96
[]
no_license
https://github.com/gitddabong/acmicpc
b1c6e162ed1954e889945622cea02e485dd6b9cc
433f5be41f29757f4dc9f6ff502950786714566a
refs/heads/master
2023-07-29T04:25:20.933055
2021-09-06T12:03:25
2021-09-06T12:03:25
394,655,049
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
import sys n = int(sys.stdin.readline()) # 자바 식 입력 array = [0] * n for i in range(n): array[i] = int(sys.stdin.readline()) for i in sorted(array): sys.stdout.write(str(i) + '\n') # 자바 식 출력
UTF-8
Python
false
false
229
py
23
2751.py
16
0.569378
0.564593
0
10
20
49
drkovalskyi/Bmm5
12,644,383,766,916
28d98425583aa0915a79a1242f0406fe8d24aef0
274169d01aaeb3704aee765135d84c9f28349787
/NanoAOD/validation/fake_muon_sources.py
760ef60ae328c6cb65582eb602d3c23a4c80b593
[]
no_license
https://github.com/drkovalskyi/Bmm5
891acb1ac50b0ad3061cf44a72a8aab01c34449d
c56b03c919a9827ca99baa46dc8a8b7b28aaf924
refs/heads/master
2023-08-31T11:24:55.148995
2023-08-29T12:35:40
2023-08-29T12:35:40
213,879,062
1
6
null
false
2021-01-08T15:30:01
2019-10-09T09:41:48
2021-01-05T14:08:25
2021-01-05T14:08:32
36,062
1
2
4
Python
false
false
import ROOT import sys, os, subprocess from DataFormats.FWLite import Events, Handle from ROOT import TFile,TTree,TH1,TROOT,TDirectory,TPad,TCanvas,TColor from math import * # output_path = "/afs/cern.ch/user/d/dmytro/www/public_html/plots/bmm5_NanoAODv6-508/muon_fake_sources_bhh_medium" # output_path = "/afs/cern.ch/user/d/dmytro/www/public_html/plots/bmm5_NanoAODv6-508/muon_fake_sources_bhh_loose" output_path = "/afs/cern.ch/user/d/dmytro/www/public_html/plots/bmm5_NanoAODv6-508/muon_fake_sources_bhh_mva" # output_path = "/afs/cern.ch/user/d/dmytro/www/public_html/plots/bmm5_NanoAODv6-508/muon_fake_sources_mu_enriched_loose" dump_info = False min_pt = 4 # muon_id = None muon_id = "SoftMvaId" # muon_id = "MediumId" files = [] # find files path = "/eos/cms/store/user/dmytro/" pds = [ # 'JpsiToMuMu_JpsiPt8_TuneCP5_13TeV-pythia8' # MuEnriched - lots of data # 'QCD_Pt-20to30_MuEnrichedPt5_TuneCP5_13TeV_pythia8/crab_MuonFakeSkim_QCD_Pt-20to30_MuEnrichedPt5_TuneCP5_13TeV_pythia8_1603447990', # 'QCD_Pt-30to50_MuEnrichedPt5_TuneCP5_13TeV_pythia8/crab_MuonFakeSkim_QCD_Pt-30to50_MuEnrichedPt5_TuneCP5_13TeV_pythia8_1603446796', # 'QCD_Pt-50to80_MuEnrichedPt5_TuneCP5_13TeV_pythia8/crab_MuonFakeSkim_QCD_Pt-50to80_MuEnrichedPt5_TuneCP5_13TeV_pythia8_1603448253', # 'QCD_Pt-80to120_MuEnrichedPt5_TuneCP5_13TeV_pythia8/crab_MuonFakeSkim_QCD_Pt-80to120_MuEnrichedPt5_TuneCP5_13TeV_pythia8_1603448316', # 'QCD_Pt-120to170_MuEnrichedPt5_TuneCP5_13TeV_pythia8/crab_MuonFakeSkim_QCD_Pt-120to170_MuEnrichedPt5_TuneCP5_13TeV_pythia8_1603448633' # # Bhh 'BdToKK_BMuonFilter_SoftQCDnonD_TuneCP5_13TeV-pythia8-evtgen/crab_MuonFakeSkim_BdToKK_BMuonFilter_SoftQCDnonD_TuneCP5_13TeV-pythia8-evtgen_1603710814', 'BdToKPi_BMuonFilter_SoftQCDnonD_TuneCP5_13TeV-pythia8-evtgen/crab_MuonFakeSkim_BdToKPi_BMuonFilter_SoftQCDnonD_TuneCP5_13TeV-pythia8-evtgen_1603711796', 'BdToPiPi_BMuonFilter_SoftQCDnonD_TuneCP5_13TeV-pythia8-evtgen/crab_MuonFakeSkim_BdToPiPi_BMuonFilter_SoftQCDnonD_TuneCP5_13TeV-pythia8-evtgen_1603711442', 'BsToKK_BMuonFilter_SoftQCDnonD_TuneCP5_13TeV-pythia8-evtgen/crab_MuonFakeSkim_BsToKK_BMuonFilter_SoftQCDnonD_TuneCP5_13TeV-pythia8-evtgen_1603711503', 'BsToKPi_BMuonFilter_SoftQCDnonD_TuneCP5_13TeV-pythia8-evtgen/crab_MuonFakeSkim_BsToKPi_BMuonFilter_SoftQCDnonD_TuneCP5_13TeV-pythia8-evtgen_1603731003', 'LambdaBToPK_BMuonFilter_SoftQCDnonD_TuneCP5_13TeV-pythia8-evtgen/crab_MuonFakeSkim_LambdaBToPK_BMuonFilter_SoftQCDnonD_TuneCP5_13TeV-pythia8-evtgen_1603711266', 'LambdaBToPPi_BMuonFilter_SoftQCDnonD_TuneCP5_13TeV-pythia8-evtgen/crab_MuonFakeSkim_LambdaBToPPi_BMuonFilter_SoftQCDnonD_TuneCP5_13TeV-pythia8-evtgen_1603710918' ] for pd in pds: for f in subprocess.check_output("find %s/%s/ -type f -name '*.root'|grep crab_MuonFakeSkim" % (path, pd), shell=True).split("\n"): if f != "": files.append(f) # break # use just the first file print "Number of files: %u" % len(files) events = Events(files) # events = Events ( # [ # # '/afs/cern.ch/work/d/dmytro/projects/RunII-NanoAODv6/src/Bmm5/NanoAOD/test/muon_fake_skim.root' # '/afs/cern.ch/work/d/dmytro/projects/RunII-NanoAODv6/src/Bmm5/NanoAOD/test/BdToKPi_BMuonFilter_SoftQCDnonD_TuneCP5_13TeV-pythia8-evtgen_RunIIAutumn18MiniAOD_muon_fake_skim.root', # '/afs/cern.ch/work/d/dmytro/projects/RunII-NanoAODv6/src/Bmm5/NanoAOD/test/LambdaBToPPi_BMuonFilter_SoftQCDnonD_TuneCP5_13TeV-pythia8-evtgen+RunIIAutumn18MiniAOD-102X_upgrade2018_realistic_v15-v2+MINIAODSIM.root' # # '/afs/cern.ch/work/d/dmytro/projects/RunII-NanoAODv6/src/Bmm5/NanoAOD/test/test.root' # ] # ) def isAncestor(a,p) : if a == p : return True for i in xrange(0,p.numberOfMothers()) : if isAncestor(a,p.mother(i)) : return True return False def find_parent(cand): # look for a parent with a different PDG id parent = cand # do a fixed depth loop to avoid endless loops for i in range(100): parent = parent.mother() if not parent: break if parent.pdgId()!=cand.pdgId(): return parent return None def isGoodMuon(muon): if not muon.isTrackerMuon(): return False if not muon.isGlobalMuon(): return False if not muon.isLooseMuon(): return False if not muon.innerTrack().quality(ROOT.reco.Track.highPurity): return False if muon.pt() < min_pt or abs(muon.eta()) > 1.4: return False if muon_id: if muon_id == "SoftMvaId": if not muon.passed(ROOT.reco.Muon.SoftMvaId): return False elif muon_id == "MediumId": if not muon.passed(ROOT.reco.Muon.CutBasedIdMedium): return False else: raise Exception("Uknown muon_id: %s" % muon_id) return True def deltaPhi(phi1,phi2): return acos(cos(phi2-phi1)) def deltaR(p1,p2): return sqrt(pow(deltaPhi(p1.phi(),p2.phi()),2)+pow(p2.eta()-p1.eta(),2)) def print_canvas(output_name_without_extention, path, canvas=ROOT.gPad): if not os.path.exists(path): os.makedirs(path) canvas.Print("%s/%s.png"%(path,output_name_without_extention)) canvas.Print("%s/%s.pdf"%(path,output_name_without_extention)) canvas.Print("%s/%s.root"%(path,output_name_without_extention)) canvas.Print("%s/%s.C"%(path,output_name_without_extention)) fake_types = { 211:'pion', 321:'kaon', 2212:'proton' } def get_sim_type(muon): if muon.simPdgId() == 0: return 'not_matched' if abs(muon.simPdgId()) == 13: if abs(muon.simMotherPdgId()) == 211: return 'pion' if abs(muon.simMotherPdgId()) == 321: return 'kaon' if abs(muon.simMotherPdgId()) == 2212: return 'proton' return 'other' sim_type_to_value = { 'not_matched':0.5, 'pion':1.5, 'kaon':2.5, 'proton':3.5, 'other':4.5 } ROOT.gROOT.SetBatch(True) handlePruned = Handle ("std::vector<reco::GenParticle>") handlePacked = Handle ("std::vector<pat::PackedGenParticle>") labelPruned = ("prunedGenParticles") labelPacked = ("packedGenParticles") muonHandle, muonLabel = Handle("std::vector<pat::Muon>"),"slimmedMuons" h_sim_match = dict() h_genid_not_matched = dict() h_sim_relative_pt = dict() h_sim_relative_pt_type_matched = dict() h_sim_relative_pt_wrong_type_matched = dict() h_sim_relative_pt_matched_to_other = dict() h_sim_decay_rho = dict() h_sim_decay_rho_type_matched = dict() h_sim_decay_rho_wrong_type_matched = dict() h_sim_decay_rho_matched_to_other = dict() h_sim_decay_rho_matched_to_other_same_pt = dict() for id,name in fake_types.items(): h_sim_match[name] = ROOT.TH1D("h_sim_match_%s" % name, "Muon matching based on simulated hits", 5, 0, 5) h_sim_match[name].GetXaxis().SetBinLabel(1, "No Match") h_sim_match[name].GetXaxis().SetBinLabel(2, "Muon from Pion") h_sim_match[name].GetXaxis().SetBinLabel(3, "Muon from Kaon") h_sim_match[name].GetXaxis().SetBinLabel(4, "Muon from Proton") h_sim_match[name].GetXaxis().SetBinLabel(5, "Other") h_sim_match[name].SetFillColor(ROOT.kMagenta) h_genid_not_matched[name] = ROOT.TH1D("h_genid_not_matched_%s" % name, "Gen |pdgId| for muons not matched by simulated hits", 350, 0, 350) h_genid_not_matched[name].SetLineColor(ROOT.kBlue) h_genid_not_matched[name].SetLineWidth(2) h_genid_not_matched[name].GetXaxis().SetTitle("|pdgId|") h_sim_relative_pt[name] = ROOT.TH1D("h_sim_relative_pt_%s" % name, "Relative Pt of sim-matched particle", 100, 0, 2) h_sim_relative_pt[name].SetLineColor(ROOT.kBlue) h_sim_relative_pt[name].SetLineWidth(2) h_sim_relative_pt[name].GetXaxis().SetTitle("Pt_{sim}/Pt_{reco}") h_sim_relative_pt_type_matched[name] = ROOT.TH1D("h_sim_relative_pt_type_matched_%s" % name, "Relative Pt of sim particle (type matched)", 100, 0, 2) h_sim_relative_pt_type_matched[name].SetLineColor(ROOT.kBlue) h_sim_relative_pt_type_matched[name].SetLineWidth(2) h_sim_relative_pt_type_matched[name].GetXaxis().SetTitle("Pt_{sim}/Pt_{reco}") h_sim_relative_pt_wrong_type_matched[name] = ROOT.TH1D("h_sim_relative_pt_wrong_type_matched_%s" % name, "Relative Pt of sim particle (wrong match)", 100, 0, 2) h_sim_relative_pt_wrong_type_matched[name].SetLineColor(ROOT.kBlue) h_sim_relative_pt_wrong_type_matched[name].SetLineWidth(2) h_sim_relative_pt_wrong_type_matched[name].GetXaxis().SetTitle("Pt_{sim}/Pt_{reco}") h_sim_relative_pt_matched_to_other[name] = ROOT.TH1D("h_sim_relative_pt_matched_to_other_%s" % name, "Relative Pt of sim particle (other match)", 100, 0, 2) h_sim_relative_pt_matched_to_other[name].SetLineColor(ROOT.kBlue) h_sim_relative_pt_matched_to_other[name].SetLineWidth(2) h_sim_relative_pt_matched_to_other[name].GetXaxis().SetTitle("Pt_{sim}/Pt_{reco}") h_sim_decay_rho[name] = ROOT.TH1D("h_sim_decay_rho_%s" % name, "Decay radius", 100, 0, 400) h_sim_decay_rho[name].SetLineColor(ROOT.kBlue) h_sim_decay_rho[name].SetLineWidth(2) h_sim_decay_rho[name].GetXaxis().SetTitle("#rho") h_sim_decay_rho_type_matched[name] = ROOT.TH1D("h_sim_decay_rho_type_matched_%s" % name, "Decay radius for type matched", 100, 0, 400) h_sim_decay_rho_type_matched[name].SetLineColor(ROOT.kBlue) h_sim_decay_rho_type_matched[name].SetLineWidth(2) h_sim_decay_rho_type_matched[name].GetXaxis().SetTitle("#rho") h_sim_decay_rho_wrong_type_matched[name] = ROOT.TH1D("h_sim_decay_rho_wrong_type_matched_%s" % name, "Decay radius for matched to wrong type", 100, 0, 400) h_sim_decay_rho_wrong_type_matched[name].SetLineColor(ROOT.kBlue) h_sim_decay_rho_wrong_type_matched[name].SetLineWidth(2) h_sim_decay_rho_wrong_type_matched[name].GetXaxis().SetTitle("#rho") h_sim_decay_rho_matched_to_other[name] = ROOT.TH1D("h_sim_decay_rho_matched_to_other_%s" % name, "Decay radius for matched to other type", 100, 0, 400) h_sim_decay_rho_matched_to_other[name].SetLineColor(ROOT.kBlue) h_sim_decay_rho_matched_to_other[name].SetLineWidth(2) h_sim_decay_rho_matched_to_other[name].GetXaxis().SetTitle("#rho") h_sim_decay_rho_matched_to_other_same_pt[name] = ROOT.TH1D("h_sim_decay_rho_matched_to_other_same_pt_%s" % name, "Decay radius for matched to 'other' type with |Pt_sim/Pt_reco-1|<0.05", 100, 0, 400) h_sim_decay_rho_matched_to_other_same_pt[name].SetLineColor(ROOT.kBlue) h_sim_decay_rho_matched_to_other_same_pt[name].SetLineWidth(2) h_sim_decay_rho_matched_to_other_same_pt[name].GetXaxis().SetTitle("#rho") # loop over events count= 0 for event in events: # muons event.getByLabel(muonLabel, muonHandle) muons = muonHandle.product() has_interesting_muon = False for muon in muons: if not isGoodMuon(muon): continue has_interesting_muon = True if not has_interesting_muon: continue # check if we have a match event.getByLabel (labelPacked, handlePacked) event.getByLabel (labelPruned, handlePruned) # get the product packed = handlePacked.product() pruned = handlePruned.product() interesting_event = False for p in pruned: if abs(p.pdgId()) not in (321, 211, 2212): continue fake_type = fake_types[abs(p.pdgId())] # mother = find_parent(p) # if not mother or abs(mother.pdgId()) != 511: continue muon = None for m in muons: if deltaR(m,p)>0.1: continue if abs(m.pt()/p.pt()-1)>0.02: continue if not isGoodMuon(m): continue muon = m break if not muon: continue h_sim_match[fake_type].Fill(sim_type_to_value[get_sim_type(muon)]) if muon.simPdgId()!=0: h_sim_relative_pt[fake_type].Fill(min(muon.simPt()/p.pt(), 1.999)) h_sim_decay_rho[fake_type].Fill(min(muon.simProdRho(), 399.999)) if fake_type == get_sim_type(muon): h_sim_relative_pt_type_matched[fake_type].Fill(min(muon.simPt()/p.pt(), 1.999)) h_sim_decay_rho_type_matched[fake_type].Fill(min(muon.simProdRho(), 399.999)) else: h_sim_relative_pt_wrong_type_matched[fake_type].Fill(min(muon.simPt()/p.pt(), 1.999)) h_sim_decay_rho_wrong_type_matched[fake_type].Fill(min(muon.simProdRho(), 399.999)) if get_sim_type(muon) == 'other': h_sim_relative_pt_matched_to_other[fake_type].Fill(min(muon.simPt()/p.pt(), 1.999)) h_sim_decay_rho_matched_to_other[fake_type].Fill(min(muon.simProdRho(), 399.999)) if abs(muon.simPt()/p.pt() - 1) < 0.05: h_sim_decay_rho_matched_to_other_same_pt[fake_type].Fill(min(muon.simProdRho(), 399.999)) else: if muon.genParticle(): h_genid_not_matched[fake_type].Fill(min(abs(muon.genParticle().pdgId()),399.999)) else: h_genid_not_matched[fake_type].Fill(0) interesting_event = True status = p.statusFlags().fromHardProcessBeforeFSR() if dump_info: print "Reco Muon pt : %s eta : %s phi: %s" % (muon.pt(), muon.eta(), muon.phi()) print "\tsim_pt: %s sim_pid: %s sim_mother_pid: %s sim_prod_rho: %f sim_prod_z: %f" % (muon.simPt(), muon.simPdgId(), muon.simMotherPdgId(), muon.simProdRho(), muon.simProdZ()) print "Gen pdgId : %s pt : %s eta : %s phi : %s status: %d" %(p.pdgId(), p.pt(), p.eta(), p.phi(), status) mother = find_parent(p) if mother: status = mother.statusFlags().fromHardProcessBeforeFSR() print "\tMother PdgId : %s pt : %s eta : %s phi : %s status: %d" %(mother.pdgId(),mother.pt(),mother.eta(),mother.phi(),status) mother = find_parent(mother) if mother: status = mother.statusFlags().fromHardProcessBeforeFSR() print "\tGrand Mother PdgId : %s pt : %s eta : %s phi : %s status: %d" %(mother.pdgId(),mother.pt(),mother.eta(),mother.phi(),status) if not interesting_event: continue count += 1 # https://github.com/cms-sw/cmssw/blob/CMSSW_7_4_X/DataFormats/HepMCCandidate/interface/GenStatusFlags.h if dump_info: print "Event dump" print "Run: %u, Event: %u" % (event.eventAuxiliary().run(),event.eventAuxiliary().event()) for p in pruned : if not abs(p.pdgId()) in [511,521,531]: continue final_b = True for dau in p.daughterRefVector(): if dau.pdgId() == -p.pdgId(): final_b = False break if not final_b: continue signature = 1 for pa in packed: mother = pa.mother(0) if mother and isAncestor(p,mother) : if pa.pdgId()!=22: signature *= pa.pdgId() # if not signature in [13*13*321*321, 13*13, 13*13*321, -13*13*321]: continue # d_p4 = ROOT.Math.LorentzVector(ROOT.Math.PxPyPzE4D(ROOT.Double))() d_p4 = ROOT.reco.Candidate.LorentzVector() rad_p4 = ROOT.reco.Candidate.LorentzVector() status = p.statusFlags().fromHardProcessBeforeFSR() print " PdgId : %s pt : %s eta : %s phi : %s status: %d" %(p.pdgId(),p.pt(),p.eta(),p.phi(),status) for dau in p.daughterRefVector(): status = dau.statusFlags().fromHardProcessBeforeFSR() print " dau PdgId : %s pt : %s eta : %s phi : %s status: %d" %(dau.pdgId(),dau.pt(),dau.eta(),dau.phi(),status) if p.mother(): status = p.mother().statusFlags().fromHardProcessBeforeFSR() print " mother PdgId : %s pt : %s eta : %s phi : %s status: %d" %(p.mother().pdgId(),p.mother().pt(),p.mother().eta(),p.mother().phi(),status) for pa in packed: mother = pa.mother(0) if mother and isAncestor(p,mother) : print " PdgId : %s pt : %s eta : %s phi : %s" %(pa.pdgId(),pa.pt(),pa.eta(),pa.phi()) d_p4 += pa.p4() if pa.pdgId()==22: rad_p4 += pa.p4() print " delta: %0.5f%%" % (100.*(p.p4()-d_p4).P()/p.p4().P()) print " radiation: %0.2f%%" % (100.*rad_p4.P()/p.p4().P()) if dump_info and count >= 100: sys.exit() c1 = TCanvas("c1", "c1", 800, 800) for id,name in fake_types.items(): h_sim_match[name].Draw() print_canvas("sim_match_%s" % name, output_path) h_genid_not_matched[name].Draw() print_canvas("h_genid_not_matched_%s" % name, output_path) h_sim_relative_pt[name].Draw() print_canvas("sim_relative_pt_%s" % name, output_path) h_sim_relative_pt_type_matched[name].Draw() print_canvas("sim_relative_pt_type_matched_%s" % name, output_path) h_sim_relative_pt_wrong_type_matched[name].Draw() print_canvas("sim_relative_pt_wrong_type_matched_%s" % name, output_path) h_sim_relative_pt_matched_to_other[name].Draw() print_canvas("sim_relative_pt_matched_to_other_%s" % name, output_path) h_sim_decay_rho[name].Draw() print_canvas("sim_decay_rho_%s" % name, output_path) h_sim_decay_rho_type_matched[name].Draw() print_canvas("h_sim_decay_rho_type_matched_%s" % name, output_path) h_sim_decay_rho_wrong_type_matched[name].Draw() print_canvas("h_sim_decay_rho_wrong_type_matched_%s" % name, output_path) h_sim_decay_rho_matched_to_other[name].Draw() print_canvas("h_sim_decay_rho_matched_to_other_%s" % name, output_path) h_sim_decay_rho_matched_to_other_same_pt[name].Draw() print_canvas("h_sim_decay_rho_matched_to_other_same_pt_%s" % name, output_path) # Local Variables: # indent-tabs-mode: 1 # tab-width: 4 # python-indent: 4 # End:
UTF-8
Python
false
false
16,681
py
117
fake_muon_sources.py
98
0.693903
0.655656
0
366
44.576503
216
cfranklin11/budget-buddy
4,423,816,356,378
dd7e2b4be78e42dcc63909ffaaff553f974e837c
e1686543c1cc1a038fb8a0f08696f45cd0522f0c
/data/scripts/data_processor.py
b39f751a62bc625cfa052a393fa2c62b25769470
[ "MIT" ]
permissive
https://github.com/cfranklin11/budget-buddy
b55bf5d1a89d8e5e5f7bb1d2e942fb70b2c871cb
dc2e05dd8012d3e3b8a011cef9bc65a912092bab
refs/heads/master
2021-01-20T12:24:53.259410
2017-06-04T09:56:31
2017-06-04T09:56:31
90,361,863
0
0
null
false
2017-06-04T09:56:31
2017-05-05T09:50:20
2017-05-27T11:21:39
2017-06-04T09:56:31
4,572
0
0
0
Jupyter Notebook
null
null
import sys import os import pandas as pd import json pd.options.mode.chained_assignment = None def list_depts(): data = get_data() depts = pd.unique(data['department_name']) return depts.tolist() def clean_data(department): data = get_data() dept_data = data[data['department_name'].str.lower() == department.lower()] dept_data.loc[:, 'measure_target'] = pd.to_numeric(dept_data['measure_target'], errors='coerce') dept_data['estimate_or_actual'] = pd.to_numeric(dept_data['estimate_or_actual'], errors='coerce').fillna(0) df = dept_data[['program_name', 'year', 'deliverable', 'estimate_or_actual']].copy() toc = df[df['deliverable'] == 'Total Output Cost'].rename(columns = {'estimate_or_actual':'total_output_cost'})[['program_name', 'year', 'total_output_cost']] df2 = dept_data.copy() df3 = df.merge(toc) df4 = df[df['deliverable'] != 'Total Output Cost'] return df4.to_dict('records') def get_data(): file_path = os.path.abspath(os.path.join(os.getcwd(), 'data/cleaned_data.csv')) data = pd.read_csv(file_path) return data def main(args): if len(args) > 1 and type(args[1]) == str: return clean_data(args[1]) else: return list_depts() print(json.dumps(main(sys.argv)))
UTF-8
Python
false
false
1,273
py
57
data_processor.py
17
0.64729
0.641005
0
40
30.825
162
ComboZhc/AsciiArt
292,057,800,320
486ae2887383657e5cedeb17a42b99a18454cb0f
3477bce623936d9cddbebdb7e4bb58e9c0a5199b
/ascii_art_painter.py
ad6c34390db69cca6143b51639fd7354ccc48877
[]
no_license
https://github.com/ComboZhc/AsciiArt
bdb89f103929a738917784aadd61d5320b89d567
27e0feadb4be5bb35b92695f2b1e265916def409
refs/heads/master
2016-09-06T18:30:45.905309
2012-09-01T08:50:28
2012-09-01T08:50:28
null
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
import Image, ImageFont, ImageDraw import tkFileDialog import sys import itertools import math import numpy src = tkFileDialog.askopenfilename() if not src.strip(): print 'no file' sys.exit() ranks = [] f = open('result.txt', 'r') for l in f: a, b = l.split(',') ranks.append((int(a), float(b))) print ranks font = ImageFont.truetype('DejaVuSansMono.ttf', 32) X_STEP = 1 Y_STEP = 2 W_FONT = 20 H_FONT = 40 try: i = Image.open(src) src_pix = i.load() src_x, src_y = i.size if src_x % X_STEP > 0: src_x -= src_x % X_STEP if src_y % Y_STEP > 0: src_y -= src_y % Y_STEP dest = Image.new('RGB', (src_x * W_FONT / X_STEP, src_y * H_FONT / Y_STEP), (255, 255, 255)) dest_d = ImageDraw.Draw(dest) #np_pix = numpy.array(src_pix) for x in range(0, src_x, X_STEP): for y in range(0, src_y, Y_STEP): rgb = [0.0, 0.0, 0.0] for r in range(0, 3): for ii in range(x, x + X_STEP): for jj in range(y, y + Y_STEP): rgb[r] += src_pix[ii, jj][r] rgb = map(lambda a: a / X_STEP / Y_STEP, rgb) gray = math.pow((math.pow(rgb[0], 2.2) * 0.2973 + math.pow(rgb[1], 2.2) * 0.6274 + math.pow(rgb[2], 2.2) * 0.0753), 1 / 2.2) dst_x = x * W_FONT / X_STEP dst_y = y * H_FONT / Y_STEP rank = int(gray / 256 * len(ranks)) s = str(chr(ranks[rank][0])) dest_d.text((dst_x, dst_y), s, font=font, fill='black') dest.save('result.jpg') except IOError, e: print e
UTF-8
Python
false
false
1,444
py
4
ascii_art_painter.py
2
0.569945
0.528393
0
55
24.290909
127
DrNolanXW/NC-Online
3,238,405,383,923
267026424ba40ae2495e8efd2871b5ae50b78644
5ae475f0cb6c8f3db0fc4f01dfaeb7e25fdf9658
/NCxJLu/apps/utils/__init__.py
917e6dca2c79c1eb48a4c4d4a6be2bf7d2782d0a
[]
no_license
https://github.com/DrNolanXW/NC-Online
fff49dcec96869163ebc7552d44d8aeab2675670
b39407e0daeac569cb78e6391513faf5c263f999
refs/heads/master
2020-05-27T21:20:25.058696
2017-03-29T09:28:49
2017-03-29T09:28:49
83,666,438
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
# -*- coding: utf-8 -*- __author__ = 'wx' __date__ = '17-2-10 下午 4:54'
UTF-8
Python
false
false
74
py
25
__init__.py
23
0.457143
0.328571
0
3
22.666667
28
lanl/OPPPY
352,187,359,517
8262c3f8fd9a1fe032099c4625fd7720048a4684
d2435243de0b83b922fbc6f527517b932328fdac
/tests/my_test_opppy_dump_parser.py
6de2207f2999d35f35e867515fa34abfc0787d41
[ "BSD-3-Clause" ]
permissive
https://github.com/lanl/OPPPY
288be84f43a051ad90a8a1756b10d89c57b624a4
c3f07c93da6969a01cda7f78d48d75af98765960
refs/heads/master
2023-04-02T16:06:25.010458
2023-03-22T16:08:25
2023-03-22T16:08:25
202,370,014
3
5
NOASSERTION
false
2023-03-22T18:04:00
2019-08-14T14:44:21
2023-02-02T21:10:04
2023-03-22T18:03:59
6,750
3
5
4
Python
false
false
# a simple example opppy dump parser import sys import re from numpy import * sys.path.append('..') from opppy.plotting_help import * class my_test_opppy_dump_parser(): ''' An example OPPPY dump parser that can extract data from a simple ASCII dump file ''' def __init__(self): # initialize the cycle parsing data print("Initializing my_test_opppy_dump_parser") def build_data_dictionary(self, filename, dump_keys=None): dump_file = open(filename,'r') lines = dump_file.readlines() data = {} keys = [] if dump_keys: keys = dump_keys else: for line in lines: if "keys" in line and not dump_keys: keys = line.strip('\n').split(' ')[1:] break data = {} for line in lines: for key in keys: if key in line and len(key) is len(line.split(':')[0]): data[key] = array(str_vector_to_float_vector(line.strip('\n').split(' ')[1:])) # build xy_verts for 2d mesh plotting example if 'x' in data and 'y' in data: xy_verts = [] for x, y in zip(data['x'], data['y']): xy_verts.append([[x-0.5,y-0.5],[x+0.5,y-0.5],[x+0.5,y+0.5],[x-0.5,y+0.5]]) data['xy_verts'] = xy_verts return data
UTF-8
Python
false
false
1,388
py
58
my_test_opppy_dump_parser.py
26
0.519452
0.505043
0
44
30.318182
98
okij98/Library
10,230,612,122,375
c000ec43b8b99b62677431b8568a25ef3cbbee1f
0c6d0955c2c52f311c8e6d957258fb9ba9489ca8
/handlers/root.py
3222187d9e01d61d923a639b9ca84f6a4a127610
[]
no_license
https://github.com/okij98/Library
d8af89e3bf1ca9c1bccc4261d2f59b348d3d4859
64d3e90f8082a8c7f5bd4863fe426f4997af6c66
refs/heads/master
2017-12-03T07:27:01.649134
2016-06-01T03:10:30
2016-06-01T03:10:30
60,082,688
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
# coding=utf-8 import tornado.web import methods.readdb as mrd class UserHandler(tornado.web.RequestHandler): def get(self): username = self.get_argument("user") user_infos = mrd.select_table(table="users",column="*",condition="username",value=username) if user_infos[0][1] == 'root': x = 1 else: x = 0 self.render("root.html", users = user_infos, x = x) class AddUser(tornado.web.RequestHandler): def get(self): user_infos = mrd.select_columns(table="users", column="username") self.render("addUser.html", users = user_infos) def post(self, *args, **kwargs): username = self.get_argument("username") password = self.get_argument("password") if mrd.add_user(table="users",username=username,password=password): self.write("add success!") else: self.write("error") class DeleteUser(tornado.web.RequestHandler): def get(self): user_infos = mrd.select_columns(table="users",column="username") self.render("deleteUser.html", users = user_infos) def post(self, *args, **kwargs): username = self.get_argument("username") if mrd.delete_user(table="users",username=username): self.write("delete success!") else: self.write("error!")
UTF-8
Python
false
false
1,349
py
11
root.py
6
0.613788
0.610082
0
39
33.615385
99
paulohrpinheiro/flask-esqueletico
4,724,464,076,093
400d270165ea8bcbf6db8f2a40586d9283e0b6a3
8e7809d38c78493a36a2ab54ebab95b540937da2
/app/__init__.py
b75db0bbf85873f0c0328055f2b4f75d9be42659
[]
no_license
https://github.com/paulohrpinheiro/flask-esqueletico
169b2794075d9fb8f149d7e4875bf22db77a2b2a
4d8454c14107ea5bee396c7217ce053987248e92
refs/heads/master
2021-01-12T12:06:48.929324
2016-10-29T18:45:05
2016-10-29T18:45:05
72,302,855
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
from flask import Flask, render_template app = Flask(__name__) app.config.from_object('config') from app.mod_root.controllers import mod_root as root_module app.register_blueprint(root_module) application = app
UTF-8
Python
false
false
214
py
5
__init__.py
3
0.771028
0.771028
0
9
22.777778
60
TECH-IN-EYE/Speech-to-text-data_collection
3,178,275,823,237
2f6faff143398d4b9e5870329bd9c2683c9d914f
af40414e861a35de3eda304c46a2ebac42e49f3e
/scripts/sendtext.py
5e553aac11e757c7dab1692db02b81d7b973e31d
[]
no_license
https://github.com/TECH-IN-EYE/Speech-to-text-data_collection
147ec806c929cf0ab8191aa26cb90c1989e769c1
7bba5778593e89548a8538905f0239bb782aee57
refs/heads/main
2023-08-07T19:24:03.939888
2021-09-15T07:02:17
2021-09-15T07:02:17
408,710,261
0
1
null
true
2021-09-21T06:24:08
2021-09-21T06:24:08
2021-09-16T00:21:54
2021-09-15T11:09:19
40,403
0
0
0
null
false
false
import pyspark import json from pyspark.sql import SparkSession from kafka import KafkaProducer from kafka.errors import KafkaError #os.environ['PYSPARK_SUBMIT_ARGS'] = '--packages com.amazonaws:aws-java-sdk-pom:1.11.538,org.apache.hadoop:hadoop-aws:2.7.3 pyspark-shell' spark = SparkSession.builder.appName('testspark').getOrCreate() producer = KafkaProducer(bootstrap_servers=["b-1.demo-cluster-1.9q7lp7.c1.kafka.eu-west-1.amazonaws.com:9092", "b-2.demo-cluster-1.9q7lp7.c1.kafka.eu-west-1.amazonaws.com:9092"],api_version = (0,10,1)) df = spark.read.csv("s3://fumbabucket/Clean_Amharic.csv") print(df.head()) for row in df.head(5): #print(json.dumps(row).encode('utf-8')) future = producer.send('group6_test',json.dumps(row).encode('utf-8')) try: record_metadata = future.get(timeout=10) except KafkaError: log.exception() pass print (record_metadata.topic)
UTF-8
Python
false
false
911
py
20
sendtext.py
9
0.720088
0.673985
0
24
36.916667
138
aethylx/intern-proj-2017
3,015,067,086,971
8c6f03d66ae18ab8ba48fc2fe01c289dd8b5f76f
be21b39697605fbf07c2ae66b9e33c56c142e3b9
/project/urls.py
ba01e8a5768b65c299b1824219b3ca43e49ac974
[]
no_license
https://github.com/aethylx/intern-proj-2017
c9ad51d774267fec94b5cc4f94a0499ba4360d38
ae3831e59deca6bd62a277a7c926c173f523c476
refs/heads/master
2018-02-08T23:28:48.508245
2016-11-14T21:53:20
2016-11-14T21:53:20
96,582,845
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
from django.conf.urls import url, include from rest_framework import routers from django.contrib import admin from django.contrib.auth import views as auth_views, forms as auth_forms from django.views.generic.edit import CreateView from project import views urlpatterns = [ # router.urls, url(r'^$', auth_views.login, name='login_page'), url(r'^projects/activities/(?P<pk>\d+)/$', views.activity_edit, name='activity_edit'), url(r'^projects/activities/assign/(?P<pk>\d+)/$', views.activity_assignment, name='activity_assignment'), url(r'^projects/resources/(?P<pk>\d+)/$', views.resource_edit, name='resource_edit'), url(r'^projects/(?P<pk>\d+)/$', views.project_detail, name='project_detail'), #url(r'^projects/(?P<slug>[^/]+)/$', views.project_detail , name='project_detail'), url(r'^projects/new/$', views.project_new, name='project_new'), url(r'^projects', views.project_list, name='project_list'), url(r'^clients/(?P<pk>\d+)/$', views.client_detail, name='client_detail'), url(r'^clients/new/$', views.client_new, name='client_new'), url(r'^clients', views.client_list, name='client_list'), url(r'^home', views.index_page, name='index_page'), url(r'^logout', views.logout_page, name='logout_page'), url(r'^register/$', views.register, name='register'), # url(r'^register/$', CreateView.as_view(template_name='registration/register.html', form_class=auth_forms.UserCreationForm, success_url='/'), name='register'), # url('^accounts/', include('django.contrib.auth.urls')), url(r'^register/success/$', views.register_success, name='register_success'), url(r'^create_activity/$', views.create_activity, name='create_activity'), url(r'^update_activity/$', views.update_activity, name='update_activity'), url(r'^update_activity_dates/$', views.update_activity_dates, name='update_activity_dates'), url(r'^delete_activity/$', views.delete_activity, name='delete_activity'), url(r'^create_link/$', views.create_link, name='create_link'), url(r'^update_link/$', views.update_link, name='update_link'), url(r'^delete_link/$', views.delete_link, name='delete_link'), url(r'^update_lmodel/$', views.update_lmodel, name='update_lmodel'), url(r'^lock_object/$', views.lock_object, name='lock_object'), url(r'^unlock_object/$', views.unlock_object, name='unlock_object'), ]
UTF-8
Python
false
false
2,379
py
34
urls.py
13
0.678016
0.678016
0
38
61.605263
164
wkeenan/HTTPi
12,412,455,485,661
2452805ba645b19675189e05763a7798f9eb5372
80107d5947c802e44f3d93d021881c57e011aeb3
/shared/lib/ipi/bluetooth.py
387a477d104d67d60c609c6841344c511f0fdb05
[]
no_license
https://github.com/wkeenan/HTTPi
e4b1befc6576798adc74eda96571c7be3c302362
0302f274ae62cc87d8f764d7b3115715eea0ff0c
refs/heads/master
2015-08-07T17:09:42.783094
2013-04-21T22:10:46
2013-04-21T22:10:46
null
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
#!python # This 'bluetooth' module name clashes with the system package 'bluetoot' so we need to resolve that... # http://stackoverflow.com/questions/4816595/how-do-i-import-a-module-whose-name-clashes-with-a-module-in-my-package from __future__ import absolute_import import os import sys import atexit from ipi.utils import iPiController current_dir = os.path.dirname(os.path.abspath(__file__)) print __name__ + ".current_dir = " + current_dir try: haveBluetooth = False import bluetooth as bt haveBluetooth = True except ImportError: haveBluetooth = False #@iPiController class BluetoothController(iPiController): def __init__(self): self.btenabled = haveBluetooth self.btConnected = False return def construct(self): if (haveBluetooth): self.btenabled = True self.btConnected = False else: self.btenabled = False self.btConnected = False atexit.register(self.btDisconnect) def btConnect(self, address, port): print "btConnect address=" + address + ", port=" + str(port) if (self.btenabled): try: self.client_socket = bt.BluetoothSocket(bt.RFCOMM) self.client_socket.connect((address, port)) self.btConnected = True self.client_socket.send(chr(0)) except bt.BluetoothError, e: exc_type, exc_obj, exc_tb = sys.exc_info() fname = os.path.split(exc_tb.tb_frame.f_code.co_filename)[1] print e print(exc_type, fname, exc_tb.tb_lineno) print "Failed to connect to address=" + address + ", port=" + str(port) self.btConnected = False def btDisconnect(self): print __name__ + "Discconect" self.btSendByte(0) def btSendByte(self, byteVal): print "btSendByte = %s" % hex(byteVal) #if (self.btenabled and not self.btconnected): # self.btConnect("00:12:05:09:90:65", 1) #TODO: this seems to be asyn and next send errors due to connect not occring if (self.btConnected): self.client_socket.send(chr(int(byteVal))) def listServices(self): nearby_devices = bt.discover_devices(lookup_names=True) print "found %d devices" % len(nearby_devices) for name, addr in nearby_devices: print " %s - %s" % (addr, name) return nearby_devices ### cherrypy config (no dependancy) _cp_config = { "/" : { 'tools.gzip.on': True } } listServices.exposed = True #listServices._cp_config = {'request.show_tracebacks': False} ###
UTF-8
Python
false
false
2,779
py
20
bluetooth.py
17
0.587621
0.579345
0
95
28.242105
137
luc3001/bfg9000
16,372,415,378,287
c8acc767216fc1d22f11f191afe3eba52f4935b8
0c452a191d35a26499abec71854f8a8cdb1efc68
/test/unit/__init__.py
8f448be0d2633a5c28cb29c363d1df7b7f5aeb1b
[ "BSD-3-Clause" ]
permissive
https://github.com/luc3001/bfg9000
fe609d64d7b605fef0ffb375873729c8cf0bd641
41452e9dd12f1a44bae68d3bf44f362d283e6802
refs/heads/master
2020-09-15T17:10:52.434281
2019-11-22T04:33:26
2019-11-22T04:33:42
null
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
import mock import ntpath import os.path import posixpath import sys from six import iteritems from .. import * from bfg9000.path import Path from bfg9000.platforms.posix import PosixPath from bfg9000.platforms.windows import WindowsPath if sys.version_info >= (3,): mock_open_name = 'builtins.open' else: mock_open_name = '__builtin__.open' # Fix the mock package's mock_open function to work with iter(); note: this is # already fixed in Python 3.7.1's unittest.mock. def mock_open(*args, **kwargs): mo = mock.mock_open(*args, **kwargs) handle = mo.return_value handle.__iter__.side_effect = lambda: iter(handle.readlines.side_effect()) return mo def skip_if_platform(platform, hide=False): return skip_pred(lambda x: x.platform_name == platform, 'not supported for platform "{}"'.format(platform), hide) def only_if_platform(platform, hide=False): return skip_pred(lambda x: x.platform_name != platform, 'only supported for platform "{}"'.format(platform), hide) class AttrDict(object): def __init__(self, **kwargs): for k, v in iteritems(kwargs): setattr(self, k, v) class CrossPlatformTestCase(TestCase): _platforms = ['linux', 'winnt', 'macos'] def __init__(self, *args, **kwargs): clear_variables = kwargs.pop('clear_variables', False) variables = kwargs.pop('variables', {}) self.platform_name = kwargs.pop('platform_name', None) TestCase.__init__(self, *args, **kwargs) if self.platform_name is None: return self.env = make_env(platform=self.platform_name, clear_variables=clear_variables, variables=variables) @property def Path(self): return self.env.host_platform.Path def shortDescription(self): return self.platform_name def parameterize(self): return parameterize_tests(self, platform_name=self._platforms) class PathTestCase(TestCase): _path_infos = [ (Path, os.path, 'native'), (PosixPath, posixpath, 'posix'), (WindowsPath, ntpath, 'windows'), ] def __init__(self, *args, **kwargs): info = kwargs.pop('path_info', None) TestCase.__init__(self, *args, **kwargs) if info is None: return self.Path, self.ospath, self._desc = info def shortDescription(self): return self._desc def parameterize(self): return parameterize_tests(self, path_info=self._path_infos)
UTF-8
Python
false
false
2,566
py
23
__init__.py
22
0.626656
0.620421
0
91
27.197802
79
cyh24/media-player
1,314,259,999,789
473278eed9b438b90b811789f2179727a748747a
1fe33ab269c3f68f6dbf7cf820c9fd0dc8d77f10
/app/dbprocess.py
b58956e6fb447e56ae5f1cf2aa8265aa2ec76abb
[]
no_license
https://github.com/cyh24/media-player
703b6ef09762e598c1b26bdfc49305415961d1f1
bb379e62b20fc5236fb30ed663f731e0c9128a01
refs/heads/master
2016-08-02T22:12:07.375269
2015-07-12T15:44:04
2015-07-12T15:44:04
38,149,730
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
#!/usr/bin/python #-*- coding:utf-8 -*- from django.db import models from models import Teacher, Comments, Video, Serial import time def curTime(): return time.strftime('%Y.%m.%d-%H:%M:%S') def add_teacher(name, info=""): t = Teacher(name=name, info=info) t.save() def add_serial(title, teacher_id, video_num, videos, money, logo_img): serial = Serial(title=title, teacher=Teacher.objects.filter(id=teacher_id)[0], video_num=video_num, money=money, logo_img=logo_img) for video in videos: video.save() serial.save() serial.videos = videos serial.save() if __name__ == "__main__": t_name = "cyh" t_info = "Handsome, cool, humor." add_teacher(t_name, t_info)
UTF-8
Python
false
false
734
py
12
dbprocess.py
4
0.623978
0.621253
0
30
23.433333
82
CrisBMoya/IEEE_Fraud_Detection
2,370,821,947,644
ee2b1e6b01bce3f0ddbf642d0f43aa43ed0862b3
fbff2c7ef32a140f5780be7133f56322f1aef37b
/IEEE_Fraud_Detection_v6.py
6b8a3836c3135e296319b62c4f1fbee2144922c0
[]
no_license
https://github.com/CrisBMoya/IEEE_Fraud_Detection
3f064f50bf70dba4d92260f9e29ff820e9b58914
7d742781993896b18ae3b6a7b61f64d9f066f8ca
refs/heads/master
2020-07-23T03:56:27.809266
2019-09-17T14:37:37
2019-09-17T14:37:37
207,438,697
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
%config Completer.use_jedi = False #Import Modules import pandas as pd import zipfile as zip import plotly as plt import numpy as np import seaborn as sns import plotly.express as px import plotly.graph_objs as go import pandas_profiling from sklearn.model_selection import train_test_split from sklearn.metrics import classification_report, confusion_matrix from sklearn.model_selection import GridSearchCV from lightgbm import LGBMClassifier from imblearn.under_sampling import EditedNearestNeighbours #Load Training data TrainTransaction=pd.read_csv(zip.ZipFile('Data/train_transaction.csv.zip').open("train_transaction.csv")) TestSet_dev=pd.read_csv(zip.ZipFile('Data/test_transaction.csv.zip').open("test_transaction.csv")) #Generate quantiles BinNum=20 TrainTransaction['QuantileAmt']=pd.qcut(x=TrainTransaction['TransactionAmt'], q=BinNum, labels=['Q'+str(X) for X in range(1,(BinNum+1))]) TestSet_dev['QuantileAmt']=pd.qcut(x=TestSet_dev['TransactionAmt'], q=BinNum, labels=['Q'+str(X) for X in range(1,(BinNum+1))]) ############################################################ ############################################################ ############################################################ ############################################################ ColumnSelect=np.asarray(["C"+str(X) for X in range(1,15)]) TempTrain=TrainTransaction[ColumnSelect] TempTrain=TempTrain.join([pd.get_dummies(data=TrainTransaction["ProductCD"]), pd.get_dummies(data=TrainTransaction["P_emaildomain"]), pd.get_dummies(data=TrainTransaction["QuantileAmt"])]) pd.get_dummies(data=TrainTransaction["P_emaildomain"]).shape pd.get_dummies(data=TrainTransaction["ProductCD"]).shape TempTrain.shape #Undersample enn = EditedNearestNeighbours() X_resampled, y_resampled = enn.fit_resample(TempTrain.iloc[1:1000:], TrainTransaction['isFraud'].iloc[1:1000:]) X_resampled #Train and test sets X_train, X_test, y_train, y_test = train_test_split(X_resampled, y_resampled, test_size=0.1, random_state=42) #Set up SDG Model with Grid Search LGBMModel=LGBMClassifier() LGBMModel.fit(X_train, y_train) #Predict Predictions=LGBMModel.predict(TempTrain) #Metrics print(confusion_matrix(y_test, Predictions)) print(classification_report(y_test, Predictions)) #Save Parameters text_file = open("Params_V6.txt", "w") text_file.write("%s\n" % confusion_matrix(y_test, Predictions)) text_file.write("%s\n" % classification_report(y_test, Predictions)) text_file.close() #Try with test TestSet_dev_Temp=TestSet_dev[ColumnSelect] TestSet_dev_Temp=TestSet_dev_Temp.join([pd.get_dummies(data=TestSet_dev["ProductCD"]), pd.get_dummies(data=TestSet_dev["P_emaildomain"]), pd.get_dummies(data=TestSet_dev["QuantileAmt"])]) TestSet_dev_Temp.columns.values TestSet_dev_Temp.drop(columns='scranton.edu', inplace=True) TempTrain.columns.values ################## #Submit predictions PredictedValues_Dev=LGBMModel.predict(TestSet_dev_Temp) #Generate file SubmitResults=pd.DataFrame(data={'TransactionID':TestSet_dev['TransactionID'], 'isFraud':PredictedValues_Dev}) SubmitResults.head() SubmitResults.to_csv(path_or_buf='SubmitResults_V6.csv',index=False) #Submit through API import os RE=True if RE==True: os.system('kaggle competitions submit -c ieee-fraud-detection -f SubmitResults_V6.csv -m "V6 Submission from API with EDA"') pass
UTF-8
Python
false
false
3,318
py
12
IEEE_Fraud_Detection_v6.py
5
0.720313
0.712176
0
83
38.987952
188
tlnguyen2018/holbertonschool-higher_level_programming
16,114,717,305,278
d56998f13dbd3431f9476901875526ac01aaab94
dc0a1a670dcbf862d45f8957419b04beffc9462d
/0x0B-python-input_output/12-student.py
d2de43079656aacd0061900bdf97f32af50aaa60
[]
no_license
https://github.com/tlnguyen2018/holbertonschool-higher_level_programming
3b4e67895e1fe70b9a727893edcfafe86d159d19
fd307b134fb298e1b9e0ef19c353f0b40ae5998c
refs/heads/master
2020-04-09T09:58:55.209093
2019-01-26T01:47:36
2019-01-26T01:47:36
160,253,304
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
#!/usr/bin/python3 """ class student like task 11, adding attribute presentation condition """ class Student: def __init__(self, first_name, last_name, age): self.first_name = first_name self.last_name = last_name self.age = age def to_json(self, attrs=None): """ Check if attr is list, not return everything """ if isinstance(attrs, list): for items in attrs: if not isinstance(items, str): return self.__dict__ return {key: self.__dict__[key] for key in attrs if key in self.__dict__} return self.__dict__
UTF-8
Python
false
false
663
py
59
12-student.py
43
0.541478
0.536953
0
24
26.625
60
tomstorey/Z80_clock
11,089,605,584,234
954cc7f5d457650b01bf06f7517cb78c6f47a89e
600b7b596918aa8b96b8b2db4aa33e7b3413fc08
/source/loader.py
cf2efc71c90b15f53c9a8a4101e8fac661581e4a
[]
no_license
https://github.com/tomstorey/Z80_clock
c2f1cb5215873b3f4094b48b041bca18a3ae9800
f041c63a229f3253705ac12fa05ea7d73a1caf6b
refs/heads/master
2022-06-20T03:21:54.396241
2022-06-06T16:09:46
2022-06-06T16:09:46
252,347,460
5
1
null
false
2022-06-06T16:09:47
2020-04-02T03:37:08
2022-05-30T21:22:10
2022-06-06T16:09:46
6,185
4
1
0
Assembly
false
false
import argparse import os import time from serial import Serial from array import array def main(): parser = argparse.ArgumentParser( description='Load application software via UART' ) parser.add_argument( '-b', '--base', dest='base', type=str, required=True, help='Base address where the software will be loaded, usually in RAM. ' 'Supply in hexadecimal format. Minimum 0x0000, maximum 0xFFFF.' ) parser.add_argument( '-e', '--exec', dest='exec', action='store_true', default=False, help='After loading, execute from base address' ) parser.add_argument( 'filename', type=str, help='Filename that should be loaded. Must be a binary file, not hex.' ) args = parser.parse_args() base = int(args.base, 16) exec = args.exec filename = args.filename if not (0 <= base <= 0xFFFF): raise ValueError( 'Base has invalid value. Minimum 0x0000, maximum 0xFFFF.' ) filesize = os.stat(filename).st_size if not (1 <= filesize <= 0xFFFF): raise ValueError( 'Size of file is invalid. Minimum 1 byte, maximum 65536.' ) if base + filesize > 0xFFFF: raise ValueError( 'Base address plus file size would cause an address overflow.' ) base_le = (base & 0xFF) << 8 | (base & 0xFF00) >> 8 filesize_le = (filesize & 0xFF) << 8 | (filesize & 0xFF00) >> 8 rom = array('B') with open(filename, 'rb') as f: rom.fromfile(f, filesize) data = [ 3, (filesize & 0xFF), (filesize & 0xFF00) >> 8, (base & 0xFF), (base & 0xFF00) >> 8 ] + list(rom) if exec is True: data += [ 4, (base & 0xFF), (base & 0xFF00) >> 8 ] data = bytes(data) ser = Serial( '/dev/cu.usbserial-FT94JP1H', baudrate=57600, timeout=1 ) if ser.in_waiting > 0: ser.read(size=ser.in_waiting) # Wait for serial loader to be available print('Waiting for serial loader availability: ', end='', flush=True) failed = 0 while True: ser.write(bytes([0x01])) ser.flush() try: if ord(ser.read(size=1)) == 64: print(' OK') break except TypeError: failed += 1 print('.', end='', flush=True) if failed == 10: print(' Failed after too many attempts') return start = time.time() print('Transferring %d bytes to 0x%4.4X: ' % (filesize, base), end='', flush=True) ser.write(data) ser.flush() failed = 0 while True: try: if ord(ser.read(size=1)) == 33: break except TypeError: failed += 1 print('.', end='', flush=True) if failed == 5: print(' Failed: transfer not acknowledged') return duration = time.time() - start print(' Done in %.3fs' % duration) failed = 0 if exec is True: print('Executing from 0x%4.4X: ' % base, end='', flush=True) while True: try: if ord(ser.read(size=1)) == 43: break except TypeError: failed += 1 print('.', end='', flush=True) if failed == 2: print(' Failed: execution not acknowledged') return print(' OK') ser.close() return if __name__ == '__main__': main()
UTF-8
Python
false
false
3,637
py
24
loader.py
7
0.509761
0.48419
0
160
21.73125
86
shivamShankhdhar/ChatPlayGround
8,134,668,086,622
809d7e35b3695648c33ee91a64052703eeac3c27
edfaac4e34bd48e3fda1d726f839933ba8ae6ccb
/accounts/views.py
2e212dbbc45e9e9d9622295ba87e32a98c9d3b75
[]
no_license
https://github.com/shivamShankhdhar/ChatPlayGround
9733a88972cb0129ef8294b6b3bcf71ed42e116a
9076bc98b4e83c4123468aabb727b9bdd759aac4
refs/heads/master
2023-02-15T21:02:22.432301
2021-01-12T10:46:12
2021-01-12T10:46:12
328,175,237
1
0
null
null
null
null
null
null
null
null
null
null
null
null
null
from django.shortcuts import render, redirect from django.http import HttpResponse, request # FORMS from accounts.forms import AccountAuthenticationForm, RegistrationForm, AccountUpdateForm # IMPORTED FOR CREATING LOGIN, LOGOUT AND AUTHENTICATING TO THE USER from django.contrib.auth import login, authenticate, logout, views # imported for BASE_URL from django.conf import settings # ACCOUNTS MODELS from accounts.models import Account # required for image cropping and saving from django.core.files.storage import default_storage from django.core.files.storage import FileSystemStorage import os import cv2 import json import base64 # import requests from django.core import files TEMP_PROFILE_IMAGE_NAME = "temp_profile_image.png" # import utils from friend.utils import get_friend_request_or_false from friend.friend_request_status import FriendRequestStatus from friend.models import FriendList, FriendRequest # views def register_view(request, *args, **kwargs): template_name = 'accounts/register.html' user = request.user if user.is_authenticated: return HttpResponse(f"You are already authenticated as {user.email}.") context = {} if request.POST: form = RegistrationForm(request.POST) if form.is_valid(): form.save() # after registerration user will automatic login email = form.cleaned_data.get('email').lower() raw_password = form.cleaned_data.get('password1') account = authenticate(email=email, password=raw_password) login(request, account) destination = get_redirect_if_exists(request) if destination: return redirect(destination) return redirect("home") else: context['registration_form'] = form return render(request, template_name, context) def logout_view(request): logout(request) return redirect('home') def login_view(request, *args, **kwargs): template_name = 'accounts/login.html' context = {} if request.user.is_authenticated: return redirect('home') destination = get_redirect_if_exists(request) if request.POST: form = AccountAuthenticationForm(request.POST) if form.is_valid(): email = request.POST['email'] password = request.POST['password'] user = authenticate(email=email, password=password) if user: login(request, user) destination = get_redirect_if_exists(request) if destination: return redirect(destination) return redirect('home') else: context['login_form'] = form return render(request, template_name, context) def get_redirect_if_exists(request): redirect = None if request.GET: if request.Get.get("next"): redirect = str(request.GET.get('next')) def account_view(request, *args, **kwargs): """ -Logic here is kidn of tricky is_self (boolean) -1: NO_REQUEST_SENT 0:THEM_SENT_TO_YOU 1:YOU_SENT_TO_THEM """ context = {} template_name = 'accounts/account.html' user_id = kwargs.get("user_id") try: account = Account.objects.get(pk=user_id) except Account.DoesNotExist: return HttpResponse("that user doesn't exist.") # if account exists if account: context['id'] = account.id context['username'] = account.username context['email'] = account.email context['profile_image'] = account.profile_image.url context['hide_email'] = account.hide_email try: friend_list = FriendList.objects.get(user = account) except FriendList.DoesNotExist: friend_list = FriendList(user = account) friend_list.save() friends = friend_list.friends.all() context['friends'] = friends # state tamplate variables is_self = True is_friend = False request_sent = FriendRequestStatus.NO_REQUEST_SENT.value friend_requests = None user = request.user if user.is_authenticated and user != account: is_self = False if friends.filter(id = user.id): is_friend = True else: is_friend = False # CASE1: Request has been sent from them to you if get_friend_request_or_false(sender = account, receiver = user) != False: request_sent = FriendRequestStatus.THEM_SENT_TO_YOU.value context["pending_friend_request_id"] = get_friend_request_or_false(sender = account, receiver = user).id #CASE 2:Request has been sent to them from you: # FriendRequestStatus.YOU_SENT_TO_THEM elif get_friend_request_or_false(sender = account, receiver = user) != False: request_sent = FriendRequestStatus.YOU_SENT_TO_THEM.value else: request_sent = FriendRequestStatus.NO_REQUEST_SENT.value elif not user.is_authenticated: is_self = False # user looking to the self profile else: try: friend_requests = FriendRequest.objects.filter(receiver = user, is_active = True) except: pass context["is_self"] = is_self context['is_friend'] = is_friend context['BASE_URL'] = settings.BASE_URL context["request_sent"] = request_sent context["friend_requests"] = friend_requests return render(request, template_name, context) # search functionality def account_search_view(request, *args, **kwargs): context = {} if request.method == "GET": search_query = request.GET.get("q") if len(search_query) > 0: search_results = Account.objects.filter(email__icontains=search_query).filter( username__icontains=search_query).distinct() user = request.user accounts = [] # [(account1, True), (account2, False), ...] for account in search_results: accounts.append((account, False)) # you have no friends yet context['accounts'] = accounts return render(request, "accounts/search_result.html", context) # views commentd """ context = {} if request.method == "GET": search_query = request.GET.get("q") if len(search_query) > 0: search_results = Account.objects.filter(email__icontains=search_query).filter( username__icontains=search_query).distinct() user = request.user accounts = [] # [(account1, True), (account2, False), ...] for account in search_results: accounts.append((account, False)) # you have no friends yet context['accounts'] = accounts return render(request, "accounts/search_results.html", context) """ # end views def edit_account_view(request, *args, **kwargs): if not request.user.is_authenticated: return redirect("login") user_id = kwargs.get("user_id") account = Account.objects.get(pk=user_id) if account.pk != request.user.pk: return HttpResponse("You cannot edit someone elses profile.") context = {} if request.POST: form = AccountUpdateForm( request.POST, request.FILES, instance=request.user) if form.is_valid(): # delete old profile image so the name preserved # account.profile_image.delete() form.save() # new_username = form.cleaned_data['username'] return redirect("accounts:view", user_id=account.pk) else: form = AccountUpdateForm(request.POST, instance=request.user, initial={ "id": account.pk, "email": account.email, "username": account.username, "profile_image": account.profile_image, "hide_email": account.hide_email, } ) context['form'] = form else: form = AccountUpdateForm( initial={ "id": account.pk, "email": account.email, "username": account.username, "profile_image": account.profile_image, "hide_email": account.hide_email, } ) context['form'] = form context['DATA_UPLOAD_MAX_MEMORY_SIZE'] = settings.DATA_UPLOAD_MAX_MEMORY_SIZE return render(request, "accounts/edit_account.html", context) def save_temp_profile_image_from_base64String(imageString, user): INCORRECT_PADDING_EXCEPTION = "Incorrect padding" try: if not os.path.exists(settings.TEMP): os.mkdir(settings.TEMP) if not os.path.exists(settings.TEMP + "/" + str(user.pk)): os.mkdir(settings.TEMP + "/" + str(user.pk)) url = os.path.join(settings.TEMP + "/" + str(user.pk), TEMP_PROFILE_IMAGE_NAME) storage = FileSystemStorage(location=url) image = base64.b64decode(imageString) with storage.open('', 'wb+') as destination: destination.write(image) destination.close() return url except Exception as e: # print("exception: " + str(e)) # workaround for an issue I found if str(e) == INCORRECT_PADDING_EXCEPTION: imageString += "=" * ((4 - len(imageString) % 4) % 4) return save_temp_profile_image_from_base64String(imageString, user) return None def crop_image(request, *args, **kwargs): payload = {} user = request.user if request.POST and user.is_authenticated: try: imageString = request.POST.get("image") url = save_temp_profile_image_from_base64String(imageString, user) img = cv2.imread(url) cropX = int(float(str(request.POST.get("cropX")))) cropY = int(float(str(request.POST.get("cropY")))) cropWidth = int(float(str(request.POST.get("cropWidth")))) cropHeight = int(float(str(request.POST.get("cropHeight")))) if cropX < 0: cropX = 0 if cropY < 0: # There is a bug with cropperjs. y can be negative. cropY = 0 crop_img = img[cropY:cropY+cropHeight, cropX:cropX+cropWidth] cv2.imwrite(url, crop_img) # delete the old image user.profile_image.delete() # Save the cropped image to user model user.profile_image.save( "profile_image.png", files.File(open(url, 'rb'))) user.save() payload['result'] = "success" payload['cropped_profile_image'] = user.profile_image.url # delete temp file os.remove(url) except Exception as e: print("exception: " + str(e)) payload['result'] = "error" payload['exception'] = str(e) return HttpResponse(json.dumps(payload), content_type="application/json")
UTF-8
Python
false
false
11,321
py
14
views.py
8
0.590319
0.587316
0
314
35.050955
124
starsfamily/PyStudio
14,671,608,306,808
85d287ce95be19e451e6fe5b7349aa540f8c4898
e6722fb2cdbb58d76d05272c8a3a7b39a1eb711d
/分行显示星期(list操作).py
4a99b6b43efca263286105222d25cd494b5b264e
[]
no_license
https://github.com/starsfamily/PyStudio
cd14a3c8e064fe1e1cd2a7736f01e022d4c9b852
3665cdb9bed57b4d47d69273aeadca68a222cbb4
refs/heads/master
2021-06-19T19:36:08.232408
2021-02-14T09:09:03
2021-02-14T09:09:03
173,943,021
2
0
null
null
null
null
null
null
null
null
null
null
null
null
null
week=['Mon.', 'Tues.', 'Wedn.', 'Thur.', 'Fri.'] weekend=['Sat.', 'Sun.'] #week.append(weekend) week.extend(weekend) for i, j in enumerate(week, 1): print(i, j) for element in week: print(element)
UTF-8
Python
false
false
207
py
40
分行显示星期(list操作).py
36
0.603865
0.599034
0
9
21.888889
48
chbicibi/vae_mnist
13,709,535,649,708
2f0fb7f532f45290d8aac61e201db071111bf6a6
8a9e3584a4236c197f83fa92fafb48628ee9b4d0
/dataset.py
e5027f496a18bb4a33f3290dc2840bda8fd2efbc
[]
no_license
https://github.com/chbicibi/vae_mnist
fc86e7072fe75ad8f5d33773c04d74c4ce8225cf
41d9e199c1670896509208b7b5f0def90d7359a6
refs/heads/master
2020-04-21T06:34:21.761502
2019-02-15T10:01:37
2019-02-15T10:01:37
169,368,961
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
import argparse import configparser import math import os import shutil import numpy as np import matplotlib.pyplot as plt import myutils as ut import common as C_ DEBUG0 = False ################################################################################ # 学習イテレータ ################################################################################ class ContainerBase(object): def __init__(self, it): self.it = it self.data = None self.len = len(it) def __len__(self): return self.len def __getitem__(self, key): if type(key) is tuple: head, *tail = key if type(head) is slice: tail = (slice(None), *tail) return np.array(self[head])[tail] elif type(key) is slice: # return [self[i] for i in range(*key.indices(self.len))] return SlicedContainer(self, *key.indices(self.len)) else: if key >= self.len: raise IndexError if not self.data: return self.get_data(key) if self.data[key] is None: self.data[key] = self.get_data(key) return self.data[key] # else: # raise TypeError def get_data(self, key): return self.it[key] class SlicedContainer(ContainerBase): def __init__(self, it, start=None, stop=None, step=None): super().__init__(it) self.start = start or 0 self.stop = stop or self.len self.step = step or 1 self.last = self.stop - (self.stop - self.start) % self.step self.len = math.ceil((self.stop - self.start) / self.step) def get_data(self, key): if 0 <= key < self.len: k = self.start + key * self.step elif -self.len <= key < 0: k = self.last + (1 + key) * self.step else: raise IndexError return self.it[k] class MemoizeMapList(ContainerBase): ''' 入力イテラブルを加工するイテラブルオブジェクト ''' def __init__(self, fn, it, name='', cache=False, cache_path=None): self.name = name self.fn = fn self.it = it self.len = len(it) if cache: self.data = [None] * self.len else: self.data = None if cache_path: abspath = os.path.abspath(cache_path) os.makedirs(abspath, exist_ok=True) self.cache_path = abspath else: self.cache_path = None def get_data(self, key): if self.cache_path: if self.name: file = f'cache_{self.name}_{key}.npy' else: file = f'cache_{key}.npy' path = os.path.join(self.cache_path, file) if os.path.isfile(path): if DEBUG0: print(f'load(cache) {key}/{self.len}', ' '*20, end='\r') return np.load(path) else: data = self.load_data(key) np.save(path, data) return data else: return self.load_data(key) def load_data(self, key): if self.fn: return self.fn(self.it[key]) else: return self.it[key] class MapChain(ContainerBase): ''' 入力イテラブルを加工するイテラブルオブジェクト 複数のイテラブルを連結 ''' def __init__(self, fn, *its, name=''): self.name = name self.fn = fn self.its = its self.lens = list(map(len, its)) self.len = sum(self.lens) self.data = None def get_data(self, key): if self.fn: return self.fn(self.point(key)) else: return self.point(key) def point(self, key): if key < 0: key += self.len for i, n in enumerate(self.lens): if key < n: return self.its[i][key] key -= n print(key, self.lens) raise IndexError ################################################################################ # データを加工(オリジナル→) # frame => (H, W, C=[u, v, p, f, w]) ################################################################################ class Formatter(object): def __init__(self, vmin, vmax): self.vmin = vmin self.vmax = vmax def __call__(self, frame): a = frame[:, :, :2] a = (a - self.vmin) / (self.vmax - self.vmin) return a.transpose(2, 0, 1) # => (H, W, C) -> (C, H, W) ################################################################################ def __test__(): pass def get_args(): ''' docstring for get_args. ''' parser = argparse.ArgumentParser() parser.add_argument('out', nargs='?', default='new_script', help='Filename of the new script') parser.add_argument('--force', '-f', action='store_true', help='Force') parser.add_argument('--test', '-t', action='store_true', help='Run as test mode') args = parser.parse_args() return args def main(): ''' docstring for main. ''' args = get_args() if args.test: __test__() return if __name__ == '__main__': main()
UTF-8
Python
false
false
5,570
py
5
dataset.py
4
0.445367
0.442783
0
207
24.173913
80
acontenti/japanize
15,745,350,129,036
b33faae75962238f39c3e63b5855ca2b7bca0ccb
5d2ea039c1b3c4bb0d62cc2c78c63799ec2685d6
/japanize_server.py
2ede5d914faa2438896cc6fe541283c70077c17b
[ "BSD-3-Clause" ]
permissive
https://github.com/acontenti/japanize
2f22bac8446bd78e7fd5ced3942c74ac044416f7
fb9a3ee5b2393bf6466e57cb110b886563489621
refs/heads/master
2023-01-24T07:22:13.113227
2020-12-01T00:23:33
2020-12-01T00:23:33
312,388,156
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
from flask import jsonify, Flask import japanize app = Flask(__name__, static_folder="static") @app.route("/", defaults={"names": None}) @app.route("/<names>") def root(names): if names: return jsonify(list(map(japanize.transform, names.split(" ")))) else: return app.send_static_file("index.html")
UTF-8
Python
false
false
327
py
10
japanize_server.py
8
0.64526
0.64526
0
14
22.357143
71
nb-spaceship/ontology-test
1,967,095,071,765
d6ce34fbda7453003bf3f9251b81891ddac58c0e
c87cae10a1f821f141fe487c04f8b7ac4f4be9d1
/test_tool/test/test_neo_param/test_main.py
a6b30e1474e162adeec994da86b955612be187f7
[]
no_license
https://github.com/nb-spaceship/ontology-test
27e0669246b03c39591be474c2044227ad0c9553
30f240064b8883f9ff978543ef63e30f1af098e6
refs/heads/master
2020-03-17T09:24:13.272297
2018-08-31T02:30:24
2018-08-31T02:30:24
133,473,590
1
0
null
null
null
null
null
null
null
null
null
null
null
null
null
# -*- coding:utf-8 -*- import re import ddt import unittest import urllib import urllib.request import json import os import sys import getopt import time import requests import subprocess import traceback sys.path.append('..') sys.path.append('../..') import utils.connect from utils.config import Config from utils.taskdata import TaskData, Task from utils.logger import LoggerInstance as logger from utils.hexstring import * from utils.error import Error from utils.parametrizedtestcase import ParametrizedTestCase from api.apimanager import API # from utils.api.commonapi import * # from utils.api.rpcapi import RPCApi # from utils.api.init_ong_ont import * # from utils.api.contractapi import * # from test_governance_api.test_api import nodeCountCheck from test_neo_param.test_config import test_config class test_neo_param_1(ParametrizedTestCase): def test_init(self): test_config.m_contract_address= API.contract().deploy_contract_full(test_config.deploy_neo, test_config.name, test_config.desc)[0] API.node().wait_gen_block() time.sleep(5) def setUp(self): logger.open("test_neo_param/" + self._testMethodName+".log",self._testMethodName) def tearDown(self): logger.close(self.result()) def test_base_001_invokeFunction(self): process = False try: (process, response) = API.contract().invoke_function(test_config.m_contract_address, "int_to_int", "", argvs = [{"type": "int","value": "0"}], node_index = 0, sleep=1) rs = response["result"]["Result"] == '01' self.ASSERT(process and rs, "") except Exception as e: traceback.print_exc() logger.print(e.args[0]) def test_abnormal_002_invokeFunction(self): process = False try: (process, response) = API.contract().invoke_function(test_config.m_contract_address, "int_to_int", "", argvs = [{"type": "int","value": "-1"}], node_index = 0, sleep=1) rs = response["result"]["Result"] == '01' self.ASSERT(process and rs, "") except Exception as e: logger.print(e.args[0]) def test_normal_003_invokeFunction(self): process = False try: (process, response) = API.contract().invoke_function(test_config.m_contract_address, "int_to_int", "", argvs = [{"type": "int","value": "65535"}], node_index = 0, sleep=1) rs = response["result"]["Result"] == '01' self.ASSERT(process and rs, "") except Exception as e: logger.print(e.args[0]) def test_abnormal_004_invokeFunction(self): process = False try: (process, response) = API.contract().invoke_function(test_config.m_contract_address, "int_to_int", "", argvs = [{"type": "int","value": "65536"}], node_index = 0, sleep=1) rs = response["result"]["Result"] == '01' self.ASSERT(process and rs, "") except Exception as e: logger.print(e.args[0]) def test_abnormal_005_invokeFunction(self): process = False try: (process, response) = API.contract().invoke_function(test_config.m_contract_address, "int_to_int", "", argvs = [{"type": "string","value": "abc"}], node_index = 0, sleep=1) rs = response["result"]["Result"] == '01' self.ASSERT(process and rs, "") except Exception as e: logger.print(e.args[0]) def test_abnormal_006_invokeFunction(self): process = False try: (process, response) = API.contract().invoke_function(test_config.m_contract_address, "int_to_int", "", argvs = [{"type": "string","value": ""}], node_index = 0, sleep=1) rs = response["result"]["Result"] == '01' self.ASSERT(process and rs, "") except Exception as e: logger.print(e.args[0]) def test_abnormal_007_invokeFunction(self): process = False try: (process, response) = API.contract().invoke_function(test_config.m_contract_address, "int_to_int", "", argvs = [{"type": "bytearray","value": "111122223333"}], node_index = 0, sleep=1) rs = response["result"]["Result"] == '01' self.ASSERT(process and rs, "") except Exception as e: logger.print(e.args[0]) def test_abnormal_008_invokeFunction(self): process = False try: (process, response) = API.contract().invoke_function(test_config.m_contract_address, "int_to_int", "", argvs = [{"type": "bool","value": "true"}], node_index = 0, sleep=1) rs = response["result"]["Result"] == '01' self.ASSERT(process and rs, "") except Exception as e: logger.print(e.args[0]) def test_abnormal_009_invokeFunction(self): process = False try: (process, response) = API.contract().invoke_function(test_config.m_contract_address, "int_to_int", "", argvs = [{"type": "bool","value": "false"}], node_index = 0, sleep=1) rs = response["result"]["Result"] == '01' self.ASSERT(process and rs, "") except Exception as e: logger.print(e.args[0]) #string to int def test_normal_010_invokeFunction(self): process = False try: (process, response) = API.contract().invoke_function(test_config.m_contract_address, "string_to_int", "", argvs = [{"type": "string","value": "qwertyuiop"}], node_index = 0, sleep=1) rs = response["result"]["Result"] == '01' self.ASSERT(process and rs, "") except Exception as e: logger.print(e.args[0]) def test_normal_011_invokeFunction(self): process = False try: (process, response) = API.contract().invoke_function(test_config.m_contract_address, "string_to_int", "", argvs = [{"type": "string","value": "~!@@#$$%^&*()_+-={}|:\"<>?;'[] \\,./"}], node_index = 0, sleep=1) rs = response["result"]["Result"] == '01' self.ASSERT(process and rs, "") except Exception as e: logger.print(e.args[0]) def test_normal_012_invokeFunction(self): process = False try: (process, response) = API.contract().invoke_function(test_config.m_contract_address, "string_to_int", "", argvs = [{"type": "string","value": "abcd1234_))()*(&^$^%#"}], node_index = 0, sleep=1) rs = response["result"]["Result"] == '01' self.ASSERT(process and rs, "") except Exception as e: logger.print(e.args[0]) def test_abnormal_013_invokeFunction(self): process = False try: (process, response) = API.contract().invoke_function(test_config.m_contract_address, "string_to_int", "", argvs = [{"type": "int","value": "123"}], node_index = 0, sleep=1) rs = response["result"]["Result"] == '01' self.ASSERT(process and rs, "") except Exception as e: logger.print(e.args[0]) def test_abnormal_014_invokeFunction(self): process = False try: (process, response) = API.contract().invoke_function(test_config.m_contract_address, "string_to_int", "", argvs = [{"type": "bytearray","value": "111122223333"}], node_index = 0, sleep=1) rs = response["result"]["Result"] == '01' self.ASSERT(process and rs, "") except Exception as e: logger.print(e.args[0]) def test_abnormal_015_invokeFunction(self): process = False try: (process, response) = API.contract().invoke_function(test_config.m_contract_address, "string_to_int", "", argvs = [{"type": "bool","value": "true"}], node_index = 0, sleep=1) rs = response["result"]["Result"] == '01' self.ASSERT(process and rs, "") except Exception as e: logger.print(e.args[0]) def test_abnormal_016_invokeFunction(self): process = False try: (process, response) = API.contract().invoke_function(test_config.m_contract_address, "string_to_int", "", argvs = [{"type": "bool","value": "false"}], node_index = 0, sleep=1) rs = response["result"]["Result"] == '01' self.ASSERT(process and rs, "") except Exception as e: logger.print(e.args[0]) def test_normal_024_invokeFunction(self): process = False try: (process, response) = API.contract().invoke_function(test_config.m_contract_address, "bool_to_int", "", argvs = [{"type": "bool","value": "true"}], node_index = 0, sleep=1) rs = response["result"]["Result"] == '01' self.ASSERT(process and rs, "") except Exception as e: logger.print(e.args[0]) def test_abnormal_025_invokeFunction(self): process = False try: (process, response) = API.contract().invoke_function(test_config.m_contract_address, "bool_to_int", "", argvs = [{"type": "int","value": "0"}], node_index = 0, sleep=1) rs = response["result"]["Result"] == '01' self.ASSERT(process and rs, "") except Exception as e: logger.print(e.args[0]) def test_normal_026_invokeFunction(self): process = False try: (process, response) = API.contract().invoke_function(test_config.m_contract_address, "bool_to_int", "", argvs = [{"type": "bool","value": "false"}], node_index = 0, sleep=1) rs = response["result"]["Result"] == '01' self.ASSERT(process and rs, "") except Exception as e: logger.print(e.args[0]) def test_abnormal_027_invokeFunction(self): process = False try: (process, response) = API.contract().invoke_function(test_config.m_contract_address, "bool_to_int", "", argvs = [{"type": "int","value": "1"}], node_index = 0, sleep=1) rs = response["result"]["Result"] == '01' self.ASSERT(process and rs, "") except Exception as e: logger.print(e.args[0]) def test_abnormal_028_invokeFunction(self): process = False try: (process, response) = API.contract().invoke_function(test_config.m_contract_address, "bool_to_int", "", argvs = [{"type": "string","value": "abc"}], node_index = 0, sleep=1) rs = response["result"]["Result"] == '01' self.ASSERT(process and rs, "") except Exception as e: logger.print(e.args[0]) def test_abnormal_029_invokeFunction(self): process = False try: (process, response) = API.contract().invoke_function(test_config.m_contract_address, "bool_to_int", "", argvs = [{"type": "string","value": ""}], node_index = 0, sleep=1) rs = response["result"]["Result"] == '01' self.ASSERT(process and rs, "") except Exception as e: logger.print(e.args[0]) def test_abnormal_030_invokeFunction(self): process = False try: (process, response) = API.contract().invoke_function(test_config.m_contract_address, "bool_to_int", "", argvs = [{"type": "bytearray","value": "111122223333"}], node_index = 0, sleep=1) rs = response["result"]["Result"] == '01' self.ASSERT(process and rs, "") except Exception as e: logger.print(e.args[0]) def test_normal_031_invokeFunction(self): process = False try: (process, response) = API.contract().invoke_function(test_config.m_contract_address, "byte_to_int", "", argvs = [{"type": "bytearray","value": "1234567890abcdef"}], node_index = 0, sleep=1) rs = response["result"]["Result"] == '01' self.ASSERT(process and rs, "") except Exception as e: logger.print(e.args[0]) def test_abnormal_032_invokeFunction(self): process = False try: (process, response) = API.contract().invoke_function(test_config.m_contract_address, "byte_to_int", "", argvs = [{"type": "bytearray","value": ""}], node_index = 0, sleep=1) rs = response["result"]["Result"] == '01' self.ASSERT(process and rs, "") except Exception as e: logger.print(e.args[0]) def test_abnormal_033_invokeFunction(self): process = False try: (process, response) = API.contract().invoke_function(test_config.m_contract_address, "byte_to_int", "", argvs = [{"type": "int","value": "123"}], node_index = 0, sleep=1) rs = response["result"]["Result"] == '01' self.ASSERT(process and rs, "") except Exception as e: logger.print(e.args[0]) def test_abnormal_034_invokeFunction(self): process = False try: (process, response) = API.contract().invoke_function(test_config.m_contract_address, "byte_to_int", "", argvs = [{"type": "string","value": "zxcvbnm!@"}], node_index = 0, sleep=1) rs = response["result"]["Result"] == '01' self.ASSERT(process and rs, "") except Exception as e: logger.print(e.args[0]) def test_abnormal_035_invokeFunction(self): process = False try: (process, response) = API.contract().invoke_function(test_config.m_contract_address, "byte_to_int", "", argvs = [{"type": "bool","value": "true"}], node_index = 0, sleep=1) rs = response["result"]["Result"] == '01' self.ASSERT(process and rs, "") except Exception as e: logger.print(e.args[0]) def test_abnormal_036_invokeFunction(self): process = False try: (process, response) = API.contract().invoke_function(test_config.m_contract_address, "byte_to_int", "", argvs = [{"type": "bool","value": "false"}], node_index = 0, sleep=1) rs = response["result"]["Result"] == '01' self.ASSERT(process and rs, "") except Exception as e: logger.print(e.args[0]) def test_normal_037_invokeFunction(self): process = False try: (process, response) = API.contract().invoke_function(test_config.m_contract_address, "test_37", "", argvs = [{"type": "string","value": ""}], node_index = 0, sleep=1) rs = response["result"]["Result"] == '00' self.ASSERT(process and rs, "") except Exception as e: logger.print(e.args[0]) def test_abnormal_038_invokeFunction(self): process = False try: (process, response) = API.contract().invoke_function(test_config.m_contract_address, "test_38", "", argvs = [{"type": "string","value": ""}], node_index = 0, sleep=1) rs = response["result"]["Result"] == 'ff' self.ASSERT(process and rs, "") except Exception as e: logger.print(e.args[0]) def test_normal_039_invokeFunction(self): process = False try: (process, response) = API.contract().invoke_function(test_config.m_contract_address, "test_39", "", argvs = [{"type": "string","value": ""}], node_index = 0, sleep=1) rs = response["result"]["Result"] == 'ffff00' self.ASSERT(process and rs, "") except Exception as e: logger.print(e.args[0]) def test_normal_046_invokeFunction(self): process = False try: (process, response) = API.contract().invoke_function(test_config.m_contract_address, "test_46", "", argvs = [{"type": "string","value": ""}], node_index = 0, sleep=1) rs = response["result"]["Result"] == '71776572747975696f70' self.ASSERT(process and rs, "") except Exception as e: logger.print(e.args[0]) def test_normal_047_invokeFunction(self): process = False try: (process, response) = API.contract().invoke_function(test_config.m_contract_address, "test_47", "", argvs = [{"type": "string","value": ""}], node_index = 0, sleep=1) rs = response["result"]["Result"] == '7e214040232424255e262a28295f2b2d3d7b7d7c3a223c3e3f3b275b5d205c2c2e2f' self.ASSERT(process and rs, "") except Exception as e: logger.print(e.args[0]) def test_normal_048_invokeFunction(self): process = False try: (process, response) = API.contract().invoke_function(test_config.m_contract_address, "test_48", "", argvs = [{"type": "string","value": ""}], node_index = 0, sleep=1) rs = response["result"]["Result"] == '61626364313233345f292928292a28265e245e2523' self.ASSERT(process and rs, "") except Exception as e: logger.print(e.args[0]) def test_normal_060_invokeFunction(self): process = False try: (process, response) = API.contract().invoke_function(test_config.m_contract_address, "test_60", "", argvs = [{"type": "string","value": ""}], node_index = 0, sleep=1) rs = response["result"]["Result"] == '01' self.ASSERT(process and rs, "") except Exception as e: logger.print(e.args[0]) def test_normal_062_invokeFunction(self): process = False try: (process, response) = API.contract().invoke_function(test_config.m_contract_address, "test_62", "", argvs = [{"type": "string","value": ""}], node_index = 0, sleep=1) rs = response["result"]["Result"] == '00' self.ASSERT(process and rs, "") except Exception as e: logger.print(e.args[0]) def test_normal_067_invokeFunction(self): process = False try: (process, response) = API.contract().invoke_function(test_config.m_contract_address, "test_67", "", argvs = [{"type": "string","value": ""}], node_index = 0, sleep=1) rs = response["result"]["Result"] == '1234567890abcdef' self.ASSERT(process and rs, "") except Exception as e: logger.print(e.args[0]) if __name__ == '__main__': unittest.main()
UTF-8
Python
false
false
15,971
py
309
test_main.py
180
0.652871
0.621564
0
412
37.76699
211
bthate/wtfbot
15,281,493,649,436
4657cd72a41a4b52fcf33f617437d18546c6b41e
426a8b9f358d59aebace72083c6c5bfa2d00ea0f
/wtf/typ.py
74f36db95007b3dd6270f4d42368b13bc5704f89
[]
no_license
https://github.com/bthate/wtfbot
5bc3790b67ef202ab037c3f62924328f533fde56
38406a1f515fefa1bfdcae7ef6bac318dcafb364
refs/heads/master
2020-11-27T06:12:14.069206
2019-12-22T00:41:46
2019-12-22T00:41:46
229,334,522
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
import importlib import types import wtf import wtf.utl def get_cls(name): try: modname, clsname = name.rsplit(".", 1) except ValueError: modname = "wtf" clsname = name mod = importlib.import_module(modname) return getattr(mod, clsname) def get_name(o): t = type(o) if t == types.ModuleType: return o.__name__ try: n = "%s.%s" % (o.__self__.__class__.__name__, o.__name__) except AttributeError: try: n = "%s.%s" % (o.__class__.__name__, o.__name__) except AttributeError: try: n = o.__class__.__name__ except AttributeError: n = o.__name__ return n def get_type(o): t = type(o) if t == type: return get_vartype(o) return str(type(o)).split()[-1][1:-2] def get_clstype(o): try: return "%s.%s" % (o.__class__.__module__, o.__class__.__name__) except AttributeError: pass def get_objtype(o): try: return "%s.%s" % (o.__self__.__module__, o.__self__.__name__) except AttributeError: pass def get_vartype(o): try: return "%s.%s" % (o.__module__, o.__name__) except AttributeError: pass
UTF-8
Python
false
false
1,241
py
34
typ.py
32
0.501209
0.497985
0
53
22.396226
71
JoaoFaria025/udemyDjango
12,034,498,378,480
b7ac39537916956acdeb510cbdbd707f8dbb0217
7784538b4c6fd2f7cc1a228ce99177f9f7e2af5c
/base/views.py
e7cdba6b7c16a76f4c7710cc975833f3dc5bf4eb
[]
no_license
https://github.com/JoaoFaria025/udemyDjango
0b8e3c2921bb98462540849503a2e81662d4b9ea
7cfdaee1f5a994b7a23ba334e4da73fa58075484
refs/heads/main
2023-07-29T05:01:37.126645
2021-08-27T01:45:37
2021-08-27T01:45:37
null
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
from django.shortcuts import render from django.contrib import messages from .models import Cliente from .forms import ContatoForm def index(request): cliente = Cliente.objects.all() form = ContatoForm() messages.success(request,'Email enviado com sucesso') context = { 'nome' : 'POO', 'cliente':cliente, 'form': form, } return render(request,'index.html',context)
UTF-8
Python
false
false
414
py
5
views.py
4
0.676329
0.676329
0
16
24.9375
57
matheusalefe/tdd-python
7,894,149,927,478
fa0333a259d9461277244a86a86ad6c23a66d391
82478001844769e35e9c940cb30a793a5bb8786a
/framework/posts/forms.py
a4ee593e7dfb175bf520168d60409fb54c4d414e
[]
no_license
https://github.com/matheusalefe/tdd-python
49399350e2ef3946fd42b021c4a978ccef8e9e16
0883fed1631fef103d5aaa906b6a8d961bd6a811
refs/heads/master
2020-09-08T09:32:56.171909
2019-12-11T06:18:44
2019-12-11T06:18:44
221,095,218
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
from django import forms from .models import Formulario class PedidoForm(forms.ModelForm): class Meta: model = Formulario fields = [ 'nome', 'email', 'data_nascimento', 'senha', 'confirmacao_senha' ]
UTF-8
Python
false
false
289
py
32
forms.py
29
0.519031
0.519031
0
13
21.307692
34
bitcoinsSG/QRL
10,806,137,763,948
992792b32692a3cd1bf9109726a019944c6ece52
8518699dfa14a328b1f0c9239a244d6202973034
/qrl/services/PeerManager.py
d7900a0f120eb1c75eccb3f77efecc90e5d6b6de
[ "MIT" ]
permissive
https://github.com/bitcoinsSG/QRL
61d43d5766e4a5d966c53ee81a49fe97a204e5b6
88960b1ebbcfaaddc2a2b83ce069d5597ec99654
refs/heads/master
2021-01-19T07:34:03.526396
2017-11-19T21:27:44
2017-11-19T21:27:44
100,640,901
2
0
null
true
2017-08-17T20:05:24
2017-08-17T20:05:24
2017-08-17T19:47:44
2017-08-16T09:50:21
4,047
0
0
0
null
null
null
import threading from time import time, sleep import grpc from qrl.core import logger from qrl.generated import qrl_pb2 class PeerMetadata(object): DISCOVERY_THRESHOLD_SECS = 20 STABILITY_THRESHOLD_SECS = 5 def __init__(self, conn_addr): self.connected_since = time() self.peers_refreshed_at = None self.conn_addr = conn_addr self.channel = grpc.insecure_channel(conn_addr) self.stub = qrl_pb2.P2PNodeStub(self.channel) self.node_info = None @property def peers_needs_refresh(self): return self.peers_refreshed_at is None or time() - self.peers_refreshed_at > self.DISCOVERY_THRESHOLD_SECS @property def is_stable(self): return (time() - self.connected_since > self.STABILITY_THRESHOLD_SECS) and self.node_info is not None class PeerManager(object): TIMEOUT_SECS = 2 REFRESH_CYCLE_SECS = 3 def __init__(self, qrlnode): self._peers = dict() self.qrlnode = qrlnode self._lock = threading.Lock() self.thread = threading.Thread(target=self._maintain_peers) self.thread.daemon = True self.thread.start() # TODO: Create a bloom filter (as a black list) to avoid frequent reconnections # TODO: Define a banning time to avoid reconnecting to certain ips @property def peer_count(self): with self._lock: return len(self._peers) @property def stable_peer_count(self): with self._lock: return sum(1 for v in self._peers.values() if v.is_stable) def add(self, addr_list): # FIXME Get new peers every time there is a new connection with self._lock: # FIXME: Limit amount of connections # FIXME: Check banning before adding for new_peer_ip in addr_list: if new_peer_ip in ['127.0.0.1', '0', '0.0.0.0']: continue new_peer_conn_addr = '{}:9009'.format(new_peer_ip) if new_peer_conn_addr not in self._peers: self._peers[new_peer_conn_addr] = PeerMetadata(new_peer_conn_addr) def remove(self, conn_list): with self._lock: for peer_conn in conn_list: self._peers.pop(peer_conn, None) def recycle(self): with self._lock: # FIXME: Flush old connections to promote change, swap peers, etc. Use hash logic pass def stable_peers(self) -> list: with self._lock: return [v for v in self._peers.values() if v.is_stable] def _all_peers(self): # FIXME: Improve this. Make a temporary copy for now with self._lock: tmp = list(self._peers.values()) return iter(tmp) def _add_peers_callback(self, response_future): if response_future.code() == grpc.StatusCode.OK: res = response_future.result() response_future.pm.node_info = res.node_info peer_list = (peer.ip for peer in res.known_peers) self.add(peer_list) # TODO: check version/network_id or remove node # res.node_info.version # res.node_info.network_id else: self.remove([response_future.pm.conn_addr]) def _update_state_callback(self, respose_future): if respose_future.code() == grpc.StatusCode.OK: res = respose_future.result() respose_future.pm.node_info = res.node_info # TODO: check version/network_id or remove node # res.node_info.version # res.node_info.network_id else: self.remove([respose_future.pm.conn_addr]) def _maintain_peers(self): while True: try: for peer_metadata in self._all_peers(): if peer_metadata.peers_needs_refresh: f = peer_metadata.stub.GetKnownPeers.future(qrl_pb2.GetKnownPeersReq(), timeout=PeerManager.TIMEOUT_SECS) f.pm = peer_metadata f.add_done_callback(self._add_peers_callback) else: f = peer_metadata.stub.GetNodeState.future(qrl_pb2.GetNodeStateReq(), timeout=PeerManager.TIMEOUT_SECS) f.pm = peer_metadata f.add_done_callback(self._update_state_callback) # FIXME: QRLNode should probably depend on this tmp = [] for peer_metadata in self.stable_peers(): addr = peer_metadata.conn_addr.split(':')[0] tmp.append(addr) #self.qrlnode.update_peer_addresses(tmp) sleep(self.REFRESH_CYCLE_SECS) self.recycle() except Exception as e: logger.exception(e)
UTF-8
Python
false
false
4,976
py
64
PeerManager.py
60
0.562098
0.556672
0
137
35.328467
114
OSTnm/tiny_mdb
16,303,695,888,482
cdd15969877945abe34771e96fb3b76431005ad0
eb90902a1d6d941019a91b2cf385085869229811
/thirdparty/dbfread/__init__.py
b6a3f76090536403f53d658c5e299cd61a9ca7b9
[ "Apache-2.0" ]
permissive
https://github.com/OSTnm/tiny_mdb
25d3d7c4e3570d43544049b74bdcbff5ec111823
4871f3bd6c3c6c47962d5d9e4c0e0070949ce6ae
refs/heads/master
2021-01-08T01:27:14.781341
2020-02-25T09:51:10
2020-02-25T09:51:10
241,872,163
3
0
null
null
null
null
null
null
null
null
null
null
null
null
null
""" Read DBF files with Python. Example: >>> from dbfread import DBF >>> for record in DBF('people.dbf'): ... print(record) OrderedDict([('NAME', 'Alice'), ('BIRTHDATE', datetime.date(1987, 3, 1))]) OrderedDict([('NAME', 'Bob'), ('BIRTHDATE', datetime.date(1980, 11, 12))]) Full documentation at http://dbfread.readthedocs.org/ """ __author__ = 'Ole Martin Bjorndalen' __email__ = 'ombdalen@gmail.com' __url__ = 'http://nerdly.info/ole/' __license__ = 'MIT' __version__ = '2.0.4' from .dbf import DBF from .deprecated_dbf import open, read from .exceptions import * from .field_parser import FieldParser, InvalidValue # Prevent splat import. __all__ = []
UTF-8
Python
true
false
685
py
15
__init__.py
11
0.640876
0.616058
0
27
24.37037
78
lastmikoi/python-project-boilerplate
558,345,782,955
470cd594f6df8c5ca5967f36434dd8451c4567e9
25848265586460ed8966bc6aa94efcd6cd220bd8
/setup.py
a9bebbbe7d44e407307a39fc3bcc241987c8d315
[ "CC0-1.0", "LicenseRef-scancode-public-domain" ]
permissive
https://github.com/lastmikoi/python-project-boilerplate
c4983d6451ebb7fa216257c7e148372ccd6ced15
b58fe4dd19d8bd80b9c9be5c47f2bf95e8664153
refs/heads/master
2020-04-09T01:08:14.291870
2018-12-01T00:18:44
2018-12-01T00:33:13
159,893,115
4
0
null
null
null
null
null
null
null
null
null
null
null
null
null
# --------------------------------License Notice---------------------------------- # Python Project Boilerplate - A boilerplate project for python packages # # Written in 2018 by Mickaël 'lastmikoi' FALCK <lastmikoi@lastmikoi.net> # # To the extent possible under law, the author(s) have dedicated all copyright # and related and neighboring rights to this software to the public domain # worldwide. This software is distributed without any warranty. # # You should have received a copy of the CC0 Public Domain Dedication along # with this software. If not, see # <http://creativecommons.org/publicdomain/zero/1.0/>. # --------------------------------License Notice---------------------------------- """Setuptools-backed setup module.""" import codecs import os import setuptools if __name__ == '__main__': ROOT_DIR = os.path.abspath(os.path.dirname(__file__)) # Use README.rst as source for setuptools.setup's long_description param with codecs.open(os.path.join(ROOT_DIR, 'README.rst'), encoding='utf-8') as f: LONG_DESCRIPTION = f.read() setuptools.setup( # Distutils parameters name='project_name', description='Placeholder description', long_description=LONG_DESCRIPTION, author="Mickaël 'lastmikoi' FALCK", author_email='lastmikoi@lastmikoi.net', url='https://github.com/lastmikoi/python-project-boilerplate/', packages=setuptools.find_packages(exclude=['tests']), classifiers=[ 'Programming Language :: Python :: 3', ], license='CC0 1.0 Universal', keywords='python project boilerplate skeleton template', # Setuptools parameters include_package_data=True, install_requires=[ ], extras_require={ 'dev': [ 'ipython>=7.2.0,<8', ], 'test': [ 'tox>=3.5.3,<4', 'pytest>=4.0.1,<5', 'pytest-mock>=1.10.0,<2', ], }, python_requires='>=3.6,<4', setup_requires=['setuptools_scm'], # setuptools_scm parameters use_scm_version=True, )
UTF-8
Python
false
false
2,189
py
3
setup.py
1
0.575674
0.561043
0
63
33.714286
82
Hybbon/era-scripts
17,600,776,016,237
676f0a7c2703175af3d3d52b362c02f7032df406
b4521a5e1b668bed9f253d222092313895753da5
/recommenders/cofactor/run_cofactor.py
8572452b65bf161a50553b45f7b6408001d7c76d
[]
no_license
https://github.com/Hybbon/era-scripts
0b3bd41e76c11305803a193e8aa9830695285a1e
405e8e77787d0ed641d9e31f5c4ac59ca7ecfd37
refs/heads/master
2021-03-19T14:25:49.913508
2017-12-12T16:46:39
2017-12-12T16:46:39
72,033,547
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
import CoF_pre_process as pre_CoF import Cofactorization as CoF import argparse import os def parse_args(): p = argparse.ArgumentParser() p.add_argument("data", type=str, help="path to the input files") p.add_argument("-o'", "--output_dir", type=str, default="", help="output directory (default: same as input file)") p.add_argument("-p", "--part", type=str, default="u1", help="partition used in the training") p.add_argument("--proc_folder", type=str, default="pro", help="folder containing the processed files used by CoFactor") p.add_argument("-v", "--no_validation", action='store_true', help="if specified, no validation folds are generated") p.add_argument("-c", "--config", type=str, default="", help="name of an override config file. (default: none)") parsed_p = p.parse_args() if parsed_p.output_dir == "": parsed_p.output_dir = parsed_p.data return parsed_p if __name__ == '__main__': args = parse_args() print "-----------------------------" if 'reeval' in args.data: print(args) args.no_validation = True pre_CoF.run(args) CoF.run(args) #pass args os.system('rm -r {0}*'.format(os.path.join(args.data,args.proc_folder,args.part)))
UTF-8
Python
false
false
1,317
py
57
run_cofactor.py
49
0.594533
0.593014
0
43
29.627907
86
tim-schilling/bettersimple
19,043,885,026,310
ce595cfac7cdf61cbe8624a54477012c9d22c617
853c6a1b6d17fd24c18875af7cc17d4750d4b9d7
/charge/urls.py
e3fd480c5957ee9c66909d67147e518ffcda75fc
[]
no_license
https://github.com/tim-schilling/bettersimple
f0db73bfac07a0879229ed335d8462ed151dbbb7
19abbb72db2998f7e83874d905570e309686e854
refs/heads/master
2019-01-02T03:58:40.167777
2013-11-01T00:31:37
2013-11-01T00:31:37
13,758,820
0
1
null
null
null
null
null
null
null
null
null
null
null
null
null
from django.conf.urls import patterns, include, url from .views import edit_credit_card, billing, past_charge, stripe_webhook, process_monthly_charges urlpatterns = patterns( 'charge.views', url(r'^(?i)creditcard', edit_credit_card, name="edit_card_card"), url(r'^(?i)billing', billing, name="billing"), url(r'^(?i)past_charge/(\d{4})/(\d{2})/$', past_charge, name='past_charge'), url( r'^(?i)api/stripe_webhook/(?P<token>([a-fA-F0-9]{8}-[a-fA-F0-9]{4}-[a-fA-F0-9]{4}-[a-fA-F0-9]{4}-[a-fA-F0-9]{12})+)/$', stripe_webhook, name='stripe_webhook'), url( r'^(?i)api/process_monthly_charges/(?P<token>([a-fA-F0-9]{8}-[a-fA-F0-9]{4}-[a-fA-F0-9]{4}-[a-fA-F0-9]{4}-[a-fA-F0-9]{12})+)/$', process_monthly_charges, name='process_monthly_charges'), )
UTF-8
Python
false
false
797
py
71
urls.py
37
0.602258
0.559598
0
15
52.133333
136
pabloduque0/handwrittingOCR
12,919,261,635,390
55c3fa3cfd889737e5fedec99dfda74f246f5e76
7ce026a15234d64eb278849a03a8f89548276373
/SVM_OCR.py
6828445a1295210ee4dc6b09799c40421611f311
[]
no_license
https://github.com/pabloduque0/handwrittingOCR
ad97b2ff063b2f697dd2a2f8da5971ee42a0dd72
e8c93d55ef24f56314092dcd70636db19d574702
refs/heads/master
2021-06-19T15:08:24.978778
2017-06-18T20:34:46
2017-06-18T20:34:46
94,634,671
1
1
null
null
null
null
null
null
null
null
null
null
null
null
null
import cv2 import numpy as np # In this test we will use SVM instead of kNN size=20 bin_n = 16 # Number of bins svm_params = dict( kernel_type = cv2.SVM_LINEAR, svm_type = cv2.SVM_C_SVC, C=2.67, gamma=5.383 ) thisFlags = cv2.WARP_INVERSE_MAP|cv2.INTER_LINEAR # First deskew image def deskew(image): myMoments = cv2.moments(image) if abs(myMoments['mu02']) < 1e-2: return image.copy() skew = myMoments['mu11']/myMoments['mu02'] M = np.float32([[1, skew, -0.5*size*skew], [0, 1, 0]]) image = cv2.warpAffine(image,M,(size, size),flags=thisFlags) return image # HOG function def hog(image): gx = cv2.Sobel(image, cv2.CV_32F, 1, 0) gy = cv2.Sobel(image, cv2.CV_32F, 0, 1) mag, ang = cv2.cartToPolar(gx, gy) bins = np.int32(bin_n*ang/(2*np.pi)) # quantizing binvalues in (0...16) bin_cells = bins[:10,:10], bins[10:,:10], bins[:10,10:], bins[10:,10:] mag_cells = mag[:10,:10], mag[10:,:10], mag[:10,10:], mag[10:,10:] hists = [np.bincount(b.ravel(), m.ravel(), bin_n) for b, m in zip(bin_cells, mag_cells)] hist = np.hstack(hists) # hist is a 64 bit vector return hist def GaussianFilter(sigma): halfSize = 3 * sigma maskSize = 2 * halfSize + 1 mat = np.ones((maskSize,maskSize)) / (float)( 2 * np.pi * (sigma**2)) xyRange = np.arange(-halfSize, halfSize+1) xx, yy = np.meshgrid(xyRange, xyRange) x2y2 = (xx**2 + yy**2) exp_part = np.exp(-(x2y2/(2.0*(sigma**2)))) mat = mat * exp_part return mat image = cv2.imread('digits.png',0) # If desired, image can be previously smoothed gaussianFilter = GaussianFilter(1) gaussianGray1 = cv2.filter2D(image, -1, gaussianFilter) cells = [np.hsplit(row,100) for row in np.vsplit(image,50)] # First half is trainData, remaining is testData train_cells = [ i[:50] for i in cells ] test_cells = [ i[50:] for i in cells] # Training data deskewedImage = [map(deskew,row) for row in train_cells] hogData = [map(hog,row) for row in deskewedImage] trainData = np.float32(hogData).reshape(-1,64) dataResponses = np.float32(np.repeat(np.arange(10),250)[:,np.newaxis]) svm = cv2.SVM() svm.train(trainData,dataResponses, params=svm_params) svm.save('svm_data.dat') # Testing data deskewedImage = [map(deskew,row) for row in test_cells] hogData = [map(hog,row) for row in deskewedImage] testData = np.float32(hogData).reshape(-1,bin_n*4) result = svm.predict_all(testData) # Checking accuracy mask = result==dataResponses correct = np.count_nonzero(mask) print correct*100.0/result.size
UTF-8
Python
false
false
2,586
py
4
SVM_OCR.py
3
0.649652
0.595901
0
84
29.797619
92
thecut/thecut-forms
1,168,231,122,062
3bb10ecda4753c17da3104ea8a2775c7f1860d42
14479931c39afa19b345bd15dd7286e0a917615b
/thecut/forms/forms.py
f39685e0932ee56601187f36db3872232eec275d
[ "Apache-2.0" ]
permissive
https://github.com/thecut/thecut-forms
d4c208a885428763ebb29345af8a2cd42e52851b
a0aefa87c67124eb033525b1d3d5dedcd0a3ed70
refs/heads/master
2021-01-17T07:58:15.822239
2017-05-23T02:34:27
2017-05-23T02:34:27
28,124,893
0
3
null
null
null
null
null
null
null
null
null
null
null
null
null
# -*- coding: utf-8 -*- from __future__ import absolute_import, unicode_literals from django import forms from thecut.forms.utils import add_css_class class EmailTypeMixin(object): """A mixin for a :py:class:`~django.forms.Form` that sets the HTML5 ``email`` input type on any child :py:class:`~django.forms.EmailField` instances.""" def __init__(self, *args, **kwargs): super(EmailTypeMixin, self).__init__(*args, **kwargs) # Set HTML5 input type for email fields for field in self.fields.values(): if isinstance(field, forms.EmailField): field.widget.input_type = 'email' class RequiredMixin(object): """A mixin for a :py:class:`~django.forms.Form` that sets the HTML5 ``required`` attribute on any child :py:class:`~django.forms.Field` instances that is required. This mixin does not apply the `required` attribute to fields using :py:class:`~django.forms.RadioSelect` and :py:class:`~django.forms.CheckboxSelectMultiple` as the HTML5 ``required`` attribute does not behave as (usually) expected on these widgets. """ required_css_class = 'required' def __init__(self, *args, **kwargs): super(RequiredMixin, self).__init__(*args, **kwargs) # Set HTML5 required attributes. Note that if we set the required # attribute on fields with certain widgets, it will cause the form to # break by requiring EVERY option to be selected. This is not possible # with the RadioSelect widget, and in most cases won't be the desired # behaviour with the CheckboxSelectMultiple widget. If it is, the # required attribute of the widget can still be set manually in the # form. for field in self.fields.values(): if field.required and not ( isinstance(field.widget, forms.CheckboxSelectMultiple) or isinstance(field.widget, forms.RadioSelect)): field.widget.attrs.update({'required': 'required'}) class MaxLengthMixin(object): """A mixin for a :py:class:`~django.forms.Form` that sets the HTML5 ``maxlength`` attribute on any child :py:class:`~django.forms.Field` instances using the :py:class:`~django.forms.Textarea` widget. A ``max_length`` must be specified on the :py:class:`~django.forms.Field`. """ def __init__(self, *args, **kwargs): super(MaxLengthMixin, self).__init__(*args, **kwargs) # HTML5 maxlength attribute for textarea for field in self.fields.values(): if isinstance(field.widget, forms.Textarea) and field.max_length: field.widget.attrs.update({'maxlength': field.max_length}) class PlaceholderMixin(object): """A mixin for a :py:class:`~django.forms.Form` that allows you to easily set the HTML5 ``placeholder`` widget on a child :py:class:`~django.forms.Field`. To add a ``placeholder`` to a :py:class:`~django.forms.Field`, specify it in a ``placeholders`` ``dict`` on the :py:class:`~django.forms.Form`'s :py:class:`~django.forms.Form.Meta` class. For example:: class MyForm(forms.Form): foo = forms.CharField() class Meta(object): placeholders = { 'foo': 'Enter some text here.' } """ class Meta(object): placeholders = {} def __init__(self, *args, **kwargs): super(PlaceholderMixin, self).__init__(*args, **kwargs) for key, value in dict(getattr(self.Meta, 'placeholders', {})).items(): self.fields[key].widget.attrs.update({'placeholder': value}) class TimeClassMixin(object): """A mixin for a :py:class:`~django.forms.Form` that adds a ``time`` CSS class on any child :py:class:`~django.forms.Field` instances using the :py:class:`~django.forms.TimeInput` widget..""" def __init__(self, *args, **kwargs): super(TimeClassMixin, self).__init__(*args, **kwargs) # HTML5 input types and attributes for field in self.fields.values(): if isinstance(field.widget, forms.TimeInput): add_css_class(field.widget, 'time') class DateClassMixin(object): """A mixin for a :py:class:`~django.forms.Form` that adds a ``date`` CSS class on any child :py:class:`~django.forms.Field` instances using the :py:class:`~django.forms.DateInput` widget..""" def __init__(self, *args, **kwargs): super(DateClassMixin, self).__init__(*args, **kwargs) # HTML5 input types and attributes for field in self.fields.values(): if isinstance(field.widget, forms.DateInput): add_css_class(field.widget, 'date') class DateTimeClassMixin(object): """A mixin for a :py:class:`~django.forms.Form` that adds a ``datetime`` CSS class on any child :py:class:`~django.forms.Field` instances using the :py:class:`~django.forms.DateTimeInput` widget..""" def __init__(self, *args, **kwargs): super(DateTimeClassMixin, self).__init__(*args, **kwargs) # HTML5 input types and attributes for field in self.fields.values(): if isinstance(field.widget, (forms.DateTimeInput, forms.SplitDateTimeWidget)): add_css_class(field.widget, 'datetime') class DateTimeTimezoneMixin(object): """A mixin for a :py:class:`~django.forms.Form` that adds ``help_text`` to any child :py:class:`~django.forms.Field` instances using the :py:class:`~django.forms.DateTimeInput` widget. This help text contains the timezone for the field's recorded data (if any)). """ def __init__(self, *args, **kwargs): super(DateTimeTimezoneMixin, self).__init__(*args, **kwargs) self._set_timezone_help_texts(data=self.initial) def _set_timezone_help_texts(self, data): for field_name, field in self.fields.items(): field_data = data.get(field_name) if field_data and isinstance(field.widget, forms.DateTimeInput): field.help_text = field_data.tzname() def clean(self, *args, **kwargs): cleaned_data = super(DateTimeTimezoneMixin, self).clean(*args, **kwargs) self._set_timezone_help_texts(data=cleaned_data) return cleaned_data class FormMixin(DateTimeClassMixin, DateClassMixin, EmailTypeMixin, MaxLengthMixin, PlaceholderMixin, RequiredMixin, TimeClassMixin): """Form mixin. Used to extend a standard Django :py:class:`~django.forms.Form` class with useful/common behaviour. """ error_css_class = 'error' label_suffix = ''
UTF-8
Python
false
false
6,774
py
16
forms.py
6
0.628137
0.626366
0
176
37.488636
81
briantoe/MachineLearning
6,682,969,144,018
a31e013bfa7c909cd189ecef4ee6ad9c3c2a38ed
75b2d28b683033e50c9ed5407412fa98748db9e4
/ps2/problem3/problem1.py
ee896f43cfade2cb376c714c0fac58ba550b7497
[]
no_license
https://github.com/briantoe/MachineLearning
361cc4934649c14a70cf6c77ed2e1791a8ef9d9e
adbe85c1b45e38796464b50d7b39f28276846683
refs/heads/master
2020-07-18T10:16:34.420312
2019-11-23T06:12:54
2019-11-23T06:12:54
206,228,354
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
import numpy as np class Question: def __init__(self, col, val): self.col = col self.val = val def match(self, inputt): # puts given input into the question and evaulates boolean value val = inputt[self.col] return val == self.val class Leaf: def __init__(self, rows): self.prediction = class_counts(rows) class DecisionNode: def __init__(self, question, true_branch, false_branch): self.question = question self.true_branch = true_branch self.false_branch = false_branch def read_data(filename): data = [] with open(filename) as file: for line in file: data_line = [item.replace('\n', '') for item in line.split(',')] data.append(data_line) # data = np.array(data) return data # labels = data[:, 0] # grab labels # dim = len(data[0]) # dimension of the data # data = data[:, 1:dim] # grab vectors # return (data, labels) def unique_values(rows, cols): return set([row[cols] for row in rows]) def find_best_split(rows): best_gain = 0 best_question = None uncertain = gini(rows) feat_num = len(rows[0] )- 1 for col in range(feat_num): values = unique_values(rows, col) for val in values: question = Question(col,val) true_rows, false_rows = partition(rows, question) if len(true_rows) == 0 or len(false_rows) == 0: # does not divide dataset continue gain = information_gain(true_rows, false_rows, uncertain) if(gain > best_gain): best_gain = gain best_question = question return best_gain, best_question def class_counts(rows): counts = {} for item in rows: label = item[0] # grab the label off of the datapoint if label not in counts: # this counts the amount of data per classifier, stores in a dict # to reference for each classifier counts[label] = 0 counts[label] += 1 return counts def information_gain(left,right, cur_uncertain): p = float(len(left))/ (len(left) + len(right)) return cur_uncertain - p * gini(left) - (1-p) * gini(right) def gini(row): # rows counts = class_counts(row) impurity = 1 for lbl in counts: prob_of_lbl = counts[lbl] / float(len(row)) impurity -= prob_of_lbl**2 return impurity def partition(rows, question): true_rows, false_rows = [], [] for row in rows: if question.match() == True: true_rows.append(row) else: false_rows.append(row) return true_rows, false_rows def build_tree(rows): gain, question = find_best_split(rows) if gain == 0: return Leaf(rows) true_rows, false_rows = partition(rows, question) true_branch = build_tree(true_rows) false_branch = build_tree(false_rows) return DecisionNode(question, true_branch, false_branch) def classify(row, node): if isinstance(row, Leaf): return node.prediction if node.question.match(row): return classify(row, node.true_branch) else: return classify(row, node.false_branch) def main(): # attribute_types = ['cap-shape', 'cap-surface', 'cap-color', 'bruises?', 'odor', 'gill-attachment', 'gill-spacing', 'gill-size', 'gill-color', 'stalk-shape', 'stalk-root', 'stalk-surface-above-ring', 'stalk-surface-below-ring', 'stalk-color-above-ring', 'stalk-color-below-ring', 'veil-type', 'veil-color', 'ring-number', 'ring-type', 'spore-print-color', 'population', 'habitat'] filename = "mush_train.data" labels_and_data = read_data(filename) # print(str(labels_and_data)) filename = "mush_test.data" test_labels_and_data = read_data(filename) print(str(test_labels_and_data)) # tree = build_tree(labels_and_data[0]) # print(classify(labels_and_data[0], tree)) if __name__ == "__main__": main()
UTF-8
Python
false
false
4,028
py
45
problem1.py
23
0.596822
0.592602
0
145
26.77931
385
bornhack/bornhack-website
14,096,082,666,200
33c5d76fb5eae9abb7c316ca0bed722c5446615e
716e1f229eab5b5087752ebf32c3370a87958798
/src/shop/migrations/0060_productcategory_weight.py
3791ddf24dddd32ff917e723cd05c669a00c9a88
[ "BSD-3-Clause" ]
permissive
https://github.com/bornhack/bornhack-website
2e8810e00f7a48a8a7e82f02f66d67edcce6e404
767deb7f58429e9162e0c2ef79be9f0f38f37ce1
refs/heads/master
2023-08-31T23:31:04.344829
2023-08-08T21:59:19
2023-08-08T21:59:19
43,702,500
9
43
BSD-3-Clause
false
2023-09-11T09:46:14
2015-10-05T17:45:55
2023-08-08T18:32:10
2023-09-11T09:46:09
177,417
7
28
93
Python
false
false
# Generated by Django 2.2.3 on 2019-12-16 22:40 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ("shop", "0059_auto_20190718_2051"), ] operations = [ migrations.AddField( model_name="productcategory", name="weight", field=models.IntegerField( default=100, help_text="Sorting weight. Heavier items sink to the bottom.", ), ), ]
UTF-8
Python
false
false
501
py
876
0060_productcategory_weight.py
545
0.558882
0.491018
0
21
22.857143
78
simplehonestwork/django-clortho
18,425,409,704,394
fe7ceda58157e9d8c4fbc51166ced98f952faca5
098cd0dcf98ad157804c60dc2b50b0e5772b9173
/clortho/auth_backends.py
35ef08556e564eb4d9d32ba23fe0397502d145d3
[]
no_license
https://github.com/simplehonestwork/django-clortho
9347575bf5c49689f078c7bc6431ef76fa5c786d
11a006b627900f4b0b70a0ea870d6256ef0da9d5
refs/heads/master
2021-01-17T22:20:15.132489
2011-11-10T21:39:37
2011-11-10T21:39:37
2,743,111
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
from datetime import datetime from django.contrib.auth.models import User from django.conf import settings import facebook from clortho.models import Facebook class FacebookBackend: def authenticate(self, cookies): APP_ID = settings.FACEBOOK_APP_ID SECRET_KEY = settings.FACEBOOK_SECRET_KEY NS = settings.FACEBOOK_USER_NAMESPACE try: access_token = facebook.get_user_from_cookie(cookies, APP_ID, SECRET_KEY) token_not_expired = datetime.fromtimestamp(float(access_token['expires'])) > datetime.now() if 'fbs_' + APP_ID in cookies and token_not_expired: graph = facebook.GraphAPI(access_token['access_token']) user_info = graph.get_object('me') try: facebook_data = Facebook.objects.get(uid=user_info['id']) return facebook_data.user except Facebook.DoesNotExist: try: email = user_info['email'] except: email = user_info['id'] + '@dummyfbemail.com' user = User.objects.create(username=NS + user_info['id'], email=email) user.first_name = user_info['first_name'] user.last_name = user_info['last_name'] user.save() # New users get an unusable password. if settings.FACEBOOK_SET_UNUSABLE_PASSWORD: user.set_unusable_password() facebook_data = Facebook(uid=user_info['id'], url=user_info['link'], user=user) facebook_data.save() return user else: return None except: return None def get_user(self, user_id): try: return User.objects.get(pk=user_id) except: return None
UTF-8
Python
false
false
2,040
py
11
auth_backends.py
8
0.511275
0.511275
0
50
39.82
103
django-stars/cfb
10,187,662,464,299
90a0fe87e219d50f322aa3049399972051541089
1f8f002b1f9bba24dad226578b3a931693850fd2
/tests/test_header.py
8d6a8c7f2afb3a47ab42964ea5136d15de18829a
[ "BSD-2-Clause" ]
permissive
https://github.com/django-stars/cfb
408c5413f0164b96395af238861c505382bcd70c
afe2a0af7ac6dd99c4c2cea730ca1c34707bbc16
refs/heads/master
2021-01-21T17:06:50.462633
2016-04-26T08:52:34
2016-04-26T08:52:34
null
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
from os import SEEK_END, SEEK_SET from six import b, BytesIO from unittest import TestCase from cfb.exceptions import MaybeDefected, ErrorDefect, FatalDefect, \ WarningDefect from cfb.header import Header class SourceMock(BytesIO, MaybeDefected): def __init__(self, value="", raise_if=ErrorDefect): super(SourceMock, self).__init__(b(value)) MaybeDefected.__init__(self, raise_if=raise_if) def append(self, data): self.write(b(data)) self.seek(0) return self def erase(self, till=0): self.seek(till, SEEK_END if till < 0 else SEEK_SET) self.truncate(self.tell()) return self class HeaderTestCase(TestCase): def test_main(self): source = SourceMock(raise_if=WarningDefect) self.assertRaises(FatalDefect, Header, source) self.assertRaises(FatalDefect, Header, source.append("12345678")) self.assertRaises( ErrorDefect, Header, source.erase().append("\xd0\xcf\x11\xe0\xa1\xb1\x1a\xe1")) self.assertRaises(ErrorDefect, Header, source.append('1' * 16)) self.assertRaises(FatalDefect, Header, source.erase(-16).append('\0' * 16)) self.assertRaises(ErrorDefect, Header, source.append("1234567890")) self.assertRaises(WarningDefect, Header, source.erase(-10).append("12\x04\x00345678")) self.assertRaises(FatalDefect, Header, source.erase(-10).append("\x3e\x00\x04\x00123456")) self.assertRaises(ErrorDefect, Header, source.erase(-6).append("\xfe\xff1234")) self.assertRaises(ErrorDefect, Header, source.erase(-4).append("\x09\x0012")) self.assertRaises( ErrorDefect, Header, source.erase(-8).append("\x03\x00\xfe\xff\x0c\x0012")) self.assertRaises(ErrorDefect, Header, source.erase(-4).append("\x09\x0012")) self.assertRaises(ErrorDefect, Header, source.erase(-2).append("\x06\x00")) self.assertRaises(ErrorDefect, Header, source.append("1" * 6)) self.assertRaises(FatalDefect, Header, source.erase(-6).append('\0' * 6)) self.assertRaises(ErrorDefect, Header, source.append("1234" + '\0' * 32)) self.assertRaises(ErrorDefect, Header, source.erase(-36).append('\0' * 36)) self.assertEqual( Header(source.erase(-20).append('\x00\x10' + '\0' * 18)).version, (3, 0x3e))
UTF-8
Python
false
false
2,669
py
15
test_header.py
13
0.581117
0.532784
0.01124
72
36.069444
77
avogis/messenger
11,244,224,431,301
af43795b4d27c96e0745d7207b5ea6d6c048b287
7d65a982c99b49d0727aa346de103196f127e8d5
/tests/tests.py
8275262bc197fcb2b19c96cc7cad3e372fec7af4
[]
no_license
https://github.com/avogis/messenger
1451c2cbec5612787a29233ab8debf4040808d02
cfe90005cdfe6df0fd6763fcc8271beaa752c41b
refs/heads/master
2020-03-28T20:05:51.218719
2017-06-20T18:53:58
2017-06-20T18:53:58
94,607,518
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
from datetime import datetime import json from messenger.api import User, Message, user_dict, messages_dict class TestModelToDict(object): def test_usr_dict(self): timestamp = datetime.utcnow() user = User('test', 'l@a', created=timestamp, last_checked=timestamp) dict_user = user_dict(user) expected_dict = { "username": 'test', "email": 'l@a', "created": str(timestamp), "last_checked": str(timestamp), } assert dict_user == expected_dict def test_message_to_dict(self): user = User('test', 'l@a',) created = datetime.utcnow() message = Message('hej', user, created=created) dict_message = messages_dict([message]) expected_dict = { message.id: { "text": 'hej', "created": str(created) }, } assert dict_message == expected_dict def test_messages_to_dict(self, sqldb): user = User('test', 'l@a') created = datetime.utcnow() message = Message('hej', user, created=created) messages = [message, message, message] expected_dict = { message.id: { "text": 'hej', "created": str(created) }, message.id: { "text": 'hej', "created": str(created) }, message.id: { "text": 'hej', "created": str(created) } } dict_message = messages_dict(messages) assert dict_message == expected_dict class TestMessageApi(object): def test_get_user_does_not_exist(self, app): response = app.get('/messages/test/1') assert response.status_code == 404 def test_get_message_does_not_exist(self, app, sqldb): user = User('test', 'test@gmail.com') sqldb.session.add(user) sqldb.session.commit() response = app.get('/messages/test/1') assert response.status_code == 404 def test_get_message(self, app, sqldb): user = User('test', 'test@gmail.com') created = datetime.utcnow() message = Message('Hi', user, created=created) sqldb.session.add(user) sqldb.session.add(message) sqldb.session.commit() response = app.get('/messages/test/1') resp = response.data.decode('utf-8') resp_dict = json.loads(resp) assert resp_dict == '{"1": {"text": "Hi", "created": "' + str(created) + '"}}' def test_delete_message_user_does_not_exist(self, app): response = app.delete('/messages/test/1') assert response.status_code == 404 def test_message_does_not_exist(self, app, sqldb): user = User('test', 'test@gmail.com') sqldb.session.add(user) sqldb.session.commit() response = app.delete('/messages/test/1') assert response.status_code == 404 def test_delete_message(self, app, sqldb): user = User('test', 'test@gmail.com') message = Message('Hi', user) sqldb.session.add(user) sqldb.session.add(message) sqldb.session.commit() assert Message.query.get(message.id) is not None response = app.delete('/messages/test/{}'.format(message.id)) assert response.status_code == 204 assert Message.query.get(message.id) is None def test_delete_message_no_such_user(self, app, sqldb): response = app.delete('/messages/test/1') assert response.status_code == 404 class TestMessages(object): def test_get_user_messages(self, app, sqldb): user = User('test', 'test@gmail.com') created = datetime.utcnow() message1 = Message('Hi', user, created=created) message2 = Message('Where are you', user, created=created) sqldb.session.add(user) sqldb.session.add(message1) sqldb.session.add(message2) sqldb.session.commit() response = app.get('/messages/test') resp = response.data.decode('utf-8') resp_dict = json.loads(resp) assert resp_dict == ('{"1": {"text": "Hi", "created": "' + str(created) + '"}, ' + '"2": {"text": "Where are you", "created": "' + str(created) + '"}}') def test_get_user_messages_does_not_exist(self, app, sqldb): response = app.get('/messages/test') assert response.status_code == 404 def test_post_new_message_for_a_user(self, app, sqldb): user = User('test', 'test@gmail.com') sqldb.session.add(user) sqldb.session.commit() message_text = 'Hello' response = app.post('/messages/test'.format(user.id), data=message_text) assert response.status_code == 201 resp = response.data.decode('utf-8') resp_dict = json.loads(resp) assert resp_dict.get("message_id") == 1 def test_post_new_message_for_a_user_does_not_exist(self, app, sqldb): message_text = 'Do you exist?' response = app.post('/messages/test', data=message_text) assert response.status_code == 404 class TestMessageOffset(object): def test_start_not_int(self, app): response = app.get('/messages/test/offset/hej/1') assert response.status_code == 404 resp = response.data.decode('utf-8') resp_dict = json.loads(resp) assert resp_dict == "Stop and start indices must be integers" def test_stop_not_int(self, app): response = app.get('/messages/test/offset/1/hej') assert response.status_code == 404 resp = response.data.decode('utf-8') resp_dict = json.loads(resp) assert resp_dict == "Stop and start indices must be integers" def test_start_bigger_than_stop(self, app): response = app.get('/messages/test/offset/5/1') assert response.status_code == 404 resp = response.data.decode('utf-8') resp_dict = json.loads(resp) assert resp_dict == "Stopp index must be greater than start index" def test_no_user(self, app): response = app.get('/messages/test/offset/1/5') assert response.status_code == 404 def test_no_messages_for_user(self, app, sqldb): user = User('test', 'test@gmail.com') sqldb.session.add(user) sqldb.session.commit() response = app.get('/messages/test/offset/1/5') resp = response.data.decode('utf-8') resp_dict = json.loads(resp) assert resp_dict == "No messages for user test" def test_get_messages(self, app, sqldb): user = User('test', 'test@gmail.com') sqldb.session.add(user) sqldb.session.commit() created = datetime.utcnow() message1 = Message('Hi1', user, created=created) sqldb.session.add(message1) message2 = Message('Hi2', user, created=created) sqldb.session.add(message2) message3 = Message('Hi3', user, created=created) sqldb.session.add(message3) sqldb.session.commit() expected_dict = { str(message1.id): { "text": 'Hi1', "created": str(created) }, str(message2.id): { "text": 'Hi2', "created": str(created) } } response = app.get('/messages/test/offset/0/2') resp = response.data.decode('utf-8') resp_dict = json.loads(resp) assert resp_dict == expected_dict expected_dict = { str(message2.id): { "text": 'Hi2', "created": str(created) }, str(message3.id): { "text": 'Hi3', "created": str(created) } } response = app.get('/messages/test/offset/1/3') resp = response.data.decode('utf-8') resp_dict = json.loads(resp) assert resp_dict == expected_dict class TestNewMessages(object): def test_user_does_not_exist(self, app): response = app.get('/new_messages/test') assert response.status_code == 404 def test_no_new_messages_for_a_user(self, app, sqldb): user = User('test', 'test@gmail.com') sqldb.session.add(user) sqldb.session.commit() response = app.get('/new_messages/test') resp = response.data.decode('utf-8') resp_dict = json.loads(resp) assert resp_dict == "No new messages for user test"
UTF-8
Python
false
false
8,522
py
7
tests.py
5
0.566299
0.555151
0
262
31.526718
98
Yash-Patel01/mysite
14,379,550,513,320
c5285c340754e0425c5901bcf9793923ca3e5170
9a0f815030ff9721f1a82ffd2498470976546324
/project/views.py
52edfdd4089a4b310602c86d359a817c90e2de50
[]
no_license
https://github.com/Yash-Patel01/mysite
838cb8904411ed7f5566d72ff87ae2b1c905427e
c415045c4b8f8cf90fa08fde36b1263a6b713eff
refs/heads/master
2022-11-26T15:16:18.540708
2020-03-28T04:51:20
2020-03-28T04:51:20
248,917,233
1
1
null
false
2022-11-22T05:24:58
2020-03-21T06:05:12
2020-03-28T04:51:23
2022-11-22T05:24:56
8,882
1
1
2
CSS
false
false
from django.shortcuts import render from .models import course, registration, contact, blog1, comment2, coursecomment from django.contrib.auth.models import User from django.http import HttpResponse import datetime import urbandictionary as ud from mysite.settings import EMAIL_HOST_USER from django.core.mail import send_mail def index(request): # word = str(input('Enter the word :')) cour = course.objects.all() # defs = ud.define(word) # # rand = ud.random() # for i in defs: # print(i.definition) return render(request, 'index.html', {'cour': cour}) def blog(request): blogg = blog1.objects.all() return render(request, "blog.html", {'blogg': blogg}) def contect1(request): return render(request, "contact.html") def course1(request, id): course_id1 = '' user_id11 = '' user_id1 = '' count1 = 'y' current_user = request.user registration1 = registration.objects.filter(user_id=current_user.id, course_id=id) registration2 = registration.objects.all() for i in registration2: user_id2 = i.user_id course_id2 = i.course_id for i in registration1: user_id11 = i.user_id # course_id1 = i.course_id # print(course_id1) # print(user_id1) if request.method == 'POST': name = request.POST['name'] email = request.POST['email'] message = request.POST['message'] id = request.POST['id'] Comment = coursecomment(name=name, email=email, message=message, idofcourse=id) Comment.save() Comment = coursecomment.objects.all() Count = coursecomment.objects.filter(idofcourse=id).count() cour = course.objects.get(id=id) return render(request, 'course.html', {'cour': cour, 'comment': Comment,'count1': count1,'count': Count, 'registration1': registration1, 'registration2': registration2, 'user_id11': user_id11}) def message(request): id1 = '' if request.method == 'POST': id1 = request.POST['id'] contact.objects.filter(id=id1).delete() mes = contact.objects.all() return render(request, 'message.html', {'mes': mes}) def contect11(request): if request.method == 'POST': name = request.POST['name'] email = request.POST['email'] subject = request.POST['subject'] message = request.POST['message'] Contect = contact(name=name, email=email, subject=subject, message=message) Contect.save() return render(request, "contact.html") def addblog(request): if request.method == 'POST': title = request.POST['title'] name = request.POST['name'] blog = request.POST['blog'] datetime.date.today() # username = request.user.get_short_name() Blog = blog1(title=title, name=name, blog=blog) Blog.save() return render(request, "addblog.html") else: return render(request, "addblog.html") def blogpost(request, blog): if request.method == 'POST': name = request.POST['name'] email = request.POST['email'] message = request.POST['message'] id = request.POST['id'] Comment = comment2(name=name, email=email, message=message, idofblog=id) Comment.save() blogg = blog1.objects.get(id=blog) Count = comment2.objects.filter(idofblog=blog).count() comment = comment2.objects.all() return render(request, 'blog-post.html', {"blog": blogg, "comment": comment, 'count': Count}) def search(request): name = '' if request.method == 'POST': name = request.POST['name'] search1 = User.objects.filter(first_name=name) return render(request, 'search.html', {'search': search1}) def register1(request, id11): id1 = '' id2 = id11 user1 = '' email = '' if request.method == 'POST': id1 = request.POST['id'] course_id = id2 status = True user1 = User.objects.filter(id=id1) for i in user1: username = i.username first_name = i.first_name last_name = i.last_name email = i.email is_staff = i.is_staff user_id = i.id if not registration.objects.filter(user_id=user_id, username=username, first_name=first_name,last_name=last_name, email=email, is_staff=is_staff, course_id=course_id, status=status).exists(): Registration = registration(user_id=user_id, username=username, first_name=first_name,last_name=last_name, email=email, is_staff=is_staff, course_id=course_id, status=status) Registration.save() current_user = request.user registration1 = registration.objects.filter(user_id=current_user.id) for i in registration1: email = i.email subject = 'Welcome to Edusite' course_details = course.objects.filter(id=id2) for i in course_details: title = i.title name = i.name start_date = i.startdate end_date = i.enddate message ='''Dear Subscriber, I am Yash Kathrotiya. You are Enrolled in Edusite Course , Please check the Enrollment status in course page . hope you will enjoye the course tutorial which will be given by Edusite ... ''' message1 = message + '\n\t Name of the course is : '+ title message2 = message1 + '\n\t Name of the Faculty is : '+ name message4 = message2 + '\n\t Starting Date is : '+ str(start_date) message5 = message4 + '\n\t Ending Date is : '+str(end_date) message3 = message5 + " \n\n\t\t***** DO NOT REPLY ON THIS EMAIL*****" recepient = email # print(message3) send_mail(subject,message3, EMAIL_HOST_USER, [recepient],fail_silently = False) print("****************************** SUCCESS ******************************") # return render(request, 'index.html') return render(request, "register1.html") else: print('course_id :'+id2) return render(request, 'register1.html', {'course_id': id2}) def dictionary(request): yash ='' word ='' if request.method == 'POST': word = request.POST['word'] yash =ud.define(word) return render(request,'dictionary.html',{'word':yash,'word1':word})
UTF-8
Python
false
false
6,408
py
20
views.py
16
0.603464
0.59176
0
187
33.26738
199
Leezr8023/try
8,581,344,661,778
1151bd00675b48ebc5b812635357b7ef480fe1ea
20cbaebf8778829b45ec8652a42ea6afd23d244c
/test.py
8cdb49328d63c9f31096a7666dc543fcf747140c
[]
no_license
https://github.com/Leezr8023/try
c52ee5e072d5319c7d9fa30ab7d442ee87771bf3
fa4d4b586476714c8f3273c75540dabc33c3f1c5
refs/heads/master
2020-12-05T07:59:19.757814
2020-02-06T02:21:41
2020-02-06T02:21:41
232,053,012
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
import sys import time from ui_new import Ui_MainWindow from Algorithm import qtoutput from matplotlib.backends.qt_compat import QtCore, QtWidgets, is_pyqt5 from PyQt5.QtGui import QFont if is_pyqt5(): from matplotlib.backends.backend_qt5agg import ( FigureCanvas, NavigationToolbar2QT as NavigationToolbar) from mythread import myThread from matplotlib.figure import Figure from const import exitFlag,data global exitFlag class ApplicationWindow(QtWidgets.QMainWindow,Ui_MainWindow): def __init__(self): super().__init__() self.setupUi(self) dynamic_canvas = FigureCanvas(Figure(figsize=(5, 3))) self.verticalLayout.addWidget(dynamic_canvas) self.addToolBar(QtCore.Qt.BottomToolBarArea, NavigationToolbar(dynamic_canvas, self)) # self.textEdit.setFont(Qfont(14))#("Consolas", 14) self.textEdit.setFont(QFont("Consolas", 14)) self.pushButton.clicked.connect(self.MeasureThread) self.pushButton_2.clicked.connect(self.output) self.pushButton_3.clicked.connect(self.opensetting) self._dynamic_ax = dynamic_canvas.figure.subplots() self._timer = dynamic_canvas.new_timer( 100, [(self._update_canvas, (), {})]) def output(self): str = "{:.2f}" str = str.format(qtoutput()) self.textEdit.setText(str) # self.textEdit.setText(str(qtoutput())) def opensetting(self): print("槽函数触发成功") pass def MeasureThread(self): print("123") self.pushButton.setEnabled(False) thread1 = myThread(1, "Thread-1", 1) # thread2 = myThread(2, "Thread-2", 1) self._timer.start() thread1.start() print(exitFlag[0]) def _update_canvas(self): global exitFlag if exitFlag[0]==1: self._timer.stop() self.pushButton.setEnabled(True) self._dynamic_ax.clear() global data self._dynamic_ax.plot(range(1600), data[...,0])#data[...,1] self._dynamic_ax.set_xlim(0, 1600) self._dynamic_ax.set_ylim(0.0008, 0.00145) self._dynamic_ax.figure.canvas.draw() # if __name__ == "__main__": # qapp = QtWidgets.QApplication(sys.argv) # app = ApplicationWindow() # app.show() # qapp.exec_()
UTF-8
Python
false
false
2,333
py
68
test.py
7
0.632169
0.608021
0
70
32.128571
69
pl77/redditPostArchiver
4,423,816,351,520
ef60654e3fdba364bec586b439319e15eb926ebe
3b58ac6e6035968f30b45f92efd67c2395649da6
/postids.py
092301fbea8eb9df265ec3dff348700a80847ece
[]
no_license
https://github.com/pl77/redditPostArchiver
c493a9da78b8596bf820c9f234f7188295e2805f
2dcac4e736883c792c6a4966554db9cae2cf84bc
refs/heads/master
2022-08-06T23:46:50.505488
2022-07-22T15:11:13
2022-07-22T15:11:13
119,421,291
67
11
null
true
2022-07-22T15:11:14
2018-01-29T18:13:30
2022-06-11T22:18:39
2022-07-22T15:11:13
143
47
9
3
Python
false
false
#!/usr/bin/env python3 import os import sys import arrow import praw import requests import yaml from prawcore.exceptions import RequestException, NotFound from requests.exceptions import HTTPError """ Customization Configuration """ # Default post_id: # username = 'GallowBoob' # Path to which to output the file # # output_file_path = './' # The Path to the stylesheet, relative to where the html file will be stored # path_to_css = 'css/style.css' """ Reddit Post Archiver By Samuel Johnson Stoever """ if len(sys.argv) == 1: print('No username was provided. Using default username.') elif len(sys.argv) > 2: print('Too Many Arguments. Using default username.') else: username = sys.argv[1] username = username.rstrip('/') if '/u/' in username: username = username.split('/u/')[1] elif '/user/' in username: username = username.split('/user/')[1] print('Processing all posts submitted by', username) cred_path = os.path.join(os.path.dirname(os.path.abspath(__file__)), 'credentials.yml') credentials = yaml.load(open(cred_path), Loader=yaml.SafeLoader) r = praw.Reddit(client_id=credentials['client_id'], client_secret=credentials['client_secret'], user_agent=credentials['user_agent']) def get_user_post_id_set(user, first_id, postcount): user_post_id_set = set() if first_id is not None: params = dict(after=first_id, count=postcount) postgenerators = user.submissions.new(params=params) else: postgenerators = user.submissions.new() try: for post in postgenerators: post_id = "{}\n".format(post.id) user_post_id_set.add(post_id) postcount += 1 if postgenerators.yielded == 100: try: first_id = postgenerators.params['after'] except KeyError: first_id = None except NotFound: print('User not found with Reddit API. Most likely deleted.') return user_post_id_set, first_id, postcount def get_reddit_submissions(reddituser): try: user = r.redditor(reddituser) except HTTPError: print('Unable to write post ids: Invalid username or login credentials') return first_id = None postcount = 0 try: user_post_id_set, first_id, postcount = get_user_post_id_set(user, first_id, postcount) except RequestException: return post_id_set = user_post_id_set subnumber = len(user_post_id_set) print("Received", subnumber, "posts from", reddituser) totalsubnumber = subnumber while subnumber > 99: try: user_post_id_set, first_id, postcount = get_user_post_id_set(user, first_id, postcount) except RequestException: break subnumber = len(user_post_id_set) totalsubnumber += subnumber post_id_set |= user_post_id_set print("Received additional", subnumber, "posts from Reddit for", reddituser, " - Total posts received so far:", totalsubnumber, "with", len(post_id_set), "in set.") return post_id_set def get_push_submissions(reddituser): push_post_id_set = set() now = int(arrow.utcnow().timestamp()) linktemplate = "https://api.pushshift.io/reddit/search/submission/?author={author}" \ "&before={timestamp}&sort=desc&size=500" url = linktemplate.format(author=reddituser, timestamp=now) rp = requests.get(url) push = rp.json() earliest = now subnumber = len(push['data']) totalsubnumber = 0 print("Received", subnumber, "pushshift.io posts from", reddituser) while subnumber > 0: totalsubnumber += subnumber itemlist = push['data'] push['data'] = list() for item in itemlist: if item['created_utc'] < earliest: earliest = item['created_utc'] post_id = "{}\n".format(item['id']) push_post_id_set.add(post_id) url = linktemplate.format(author=reddituser, timestamp=earliest) rp = requests.get(url) push = rp.json() subnumber = len(push['data']) print("Received additional", subnumber, "posts from", reddituser, " - Total posts received so far:", totalsubnumber, "with", len(push_post_id_set), "in pushshift.io set.") return push_post_id_set def main(): reddit_post_id_set = get_reddit_submissions(username) push_post_id_set = get_push_submissions(username) post_id_set = reddit_post_id_set.union(push_post_id_set) print("Total posts submitted by", username, "in set:", len(post_id_set)) filedate = int(arrow.now().timestamp()) basedir = "/rpa" if os.environ.get('DOCKER', '0') == '1' else '.' output_file_path = "{basedir}/{username}_{timestamp}.txt".format(basedir=basedir, username=username, timestamp=filedate) with open(output_file_path, 'w', encoding='UTF-8') as post_file: post_file.writelines(post_id_set) if __name__ == '__main__': main()
UTF-8
Python
false
false
5,017
py
15
postids.py
9
0.635041
0.630855
0
147
33.129252
124
frankliumin/MemeRep
1,297,080,147,461
24bab2a1655c6c89a14221762871f36af94c3f1e
0b0af4c236f8f25dc54916bfe79dcc60323bea09
/MemeRep/cal_fit.py
711899233ed093ed05e8f92016632a9e58879bff
[]
no_license
https://github.com/frankliumin/MemeRep
2d48f13103ee31215e4c04d0f553417458353203
ecee6e882870cdcad976f427c7ef3c030de0a21d
refs/heads/master
2022-11-15T07:53:41.905472
2020-07-07T08:58:21
2020-07-07T08:58:21
null
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
# -*- coding: utf-8 -*- """ Created on Tue Jul 7 16:18:25 2020 @author: Administrator """ import numpy as np from sklearn.cluster import KMeans import random import itertools import time def cal_fit(x1, adjData): lamda = 0.5 x = x1[0:len(x1)] k = len(list(set(x))) label = list(set(x)) KKM = 0 RC = 0 for i in range(k): numi = list(x).count(label[i]) temp = [index1 for index1 in range(len(x)) if x[index1] == label[i]] # 找到类别为label[i]的下标 all_com = list(itertools.combinations(temp, 2)) # 找出该社团内所有点的下标再进行组合 L1 = 0 for index2 in all_com: if (adjData[index2[0]][index2[1]] == 1): L1 = L1 + 1 degree_i = 0 for j in temp: degree_i = degree_i + sum(adjData[j]) L2 = degree_i - 2 * L1 RC = RC + L2 / numi KKM = KKM - L1 / numi fit = -(2 * lamda * KKM + 2 * (1 - lamda) * RC) return fit
UTF-8
Python
false
false
994
py
10
cal_fit.py
8
0.532839
0.486229
0
39
23.230769
96
KalePrajwal/PythonDB
8,220,567,448,649
22af280f97ae5175e7fe03be44e6f19b51a8c7ff
307ea510630d841f1ac44ebdeab43bb1879810d7
/Search_Book.py
4d2b1751d82ed6d6185dc21bcf04b8f3859fb115
[]
no_license
https://github.com/KalePrajwal/PythonDB
06526919d6b2f7b9c1d1a9a11f2ff868eb01f296
663e5a346f04e00c62f0f4113f40fec6765991c0
refs/heads/main
2023-09-04T12:14:14.474174
2021-11-08T14:42:37
2021-11-08T14:42:37
425,726,238
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
import mysql.connector as my con = my.connect(host='bplx7zjluj8ztqdment0-mysql.services.clever-cloud.com', user='utdc4xugroxopi4q', password='l3A4aV1qVd3bMPBHITBG', database='bplx7zjluj8ztqdment0') curs=con.cursor() try: id = int(input('Enter Book id : ')) curs.execute("Select * from Books where Book_Id = %d" %id) data = curs.fetchall() if data: print() print('Book Id : %d' %id) print('Book Name : %s' %data[0][1]) print('Category : %s' %data[0][2]) print('Author : %s' %data[0][3]) print('Publication : %s' %data[0][4]) print('Edition : %s' %data[0][5]) print('Price : %.2f' %data[0][6]) else: print() print('Book not found') except: print() print('Invalid input') print() con.close()
UTF-8
Python
false
false
851
py
8
Search_Book.py
7
0.547591
0.518214
0
28
28.464286
168
fiskercui/testlanguage
4,587,025,110,945
15444854f40d9a9e1ed887769ad68895895b2f82
bb72b975267b12fb678248ce565f3fd9bd6153ee
/testpython/testjson.py
a6274232e1d4815b4b70ac3e229f674bd21943f7
[]
no_license
https://github.com/fiskercui/testlanguage
cfbcc84ffaa31a535a7f898d6c7b42dcbf2dc71b
b15746d9fa387172b749e54a90e8c6251750f192
refs/heads/master
2021-05-30T13:26:11.792822
2016-01-13T02:03:26
2016-01-13T02:03:26
null
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
import json print json.dumps(['foo', {'bar': ('baz', None, 1.0, 2)}]) print json.dumps("\"foo\bar") print json.dumps(u'\u1234') print json.dumps('\\') print json.dumps({"c": 0, "b": 0, "a": 0}, sort_keys=True) {"a": 0, "b": 0, "c": 0} from StringIO import StringIO io = StringIO() json.dump(['streaming API'], io) print "hello world" print json.JSONEncoder().encode({"foo": ["bar", "baz"]}) print io.getvalue() print("----------") io = StringIO('["streaming API"]') print json.load(io) import types import urllib2 import json duan = "--------------------------" def registerUrl(): try: url ="http://m.weather.com.cn/data/101010100.html" data = urllib2.urlopen(url).read() return data except Exception,e: print e def jsonFile(fileData): file = open("d:\json.txt","w") file.write(fileData) file.close() def praserJsonFile(jsonData): value = json.loads(jsonData) rootlist = value.keys() print rootlist print duan for rootkey in rootlist: print rootkey print duan subvalue = value[rootkey] print subvalue print duan for subkey in subvalue: print subkey,subvalue[subkey] if __name__ == "__main__": # xinput = raw_input() # x = 130 # xvalue = cmp(x,xinput) # print xvalue # print x/100.0 # data = open('test.json', 'r') # print json.load(data) with open("test.json") as json_file: json_data = json.load(json_file) print(json_data) # data = [] # with open('test.json') as f: # for line in f: # data.append(json.loads(line)) # from pprint import pprint # json_data=open('./test.json') # print json_data # data = json.load(json_data) # pprint(data) # json_data.close() # print data # data = registerUrl() # jsonFile(data) # praserJsonFile(data)
UTF-8
Python
false
false
1,722
py
101
testjson.py
44
0.633566
0.615563
0
83
19.710843
58
kurtejung/PurdueForest
16,999,480,565,501
eaacd7b6b7b06b152ca3e694521a5c77ba2f95c0
21fb99baaaeaed674cd99fb1a5d39ea54a1db827
/HLTrigger/HLTanalyzers/python/HLTopen_cff.py
69902c047a0726383b4e84bb8ec7e023b7c5985c
[]
no_license
https://github.com/kurtejung/PurdueForest
e413d4bc953df386a50d01ca09d1b648d07a926c
9c7de9e7452b0837a872cfdd428244cb46b55322
refs/heads/master
2020-05-16T20:32:30.896495
2014-06-02T18:42:18
2014-06-02T18:42:18
null
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
import FWCore.ParameterSet.Config as cms # import the whole HLT menu from HLTrigger.HLTanalyzers.HLT_FULL_cff import * hltL1IsoR9shape = cms.EDProducer( "EgammaHLTR9Producer", recoEcalCandidateProducer = cms.InputTag( "hltL1IsoRecoEcalCandidate" ), ecalRechitEB = cms.InputTag( 'hltEcalRegionalEgammaRecHit','EcalRecHitsEB' ), ecalRechitEE = cms.InputTag( 'hltEcalRegionalEgammaRecHit','EcalRecHitsEE' ), useSwissCross = cms.bool( False ) ) hltL1NonIsoR9shape = cms.EDProducer( "EgammaHLTR9Producer", recoEcalCandidateProducer = cms.InputTag( "hltL1NonIsoRecoEcalCandidate" ), ecalRechitEB = cms.InputTag( 'hltEcalRegionalEgammaRecHit','EcalRecHitsEB' ), ecalRechitEE = cms.InputTag( 'hltEcalRegionalEgammaRecHit','EcalRecHitsEE' ), useSwissCross = cms.bool( False ) ) HLTEgammaR9ShapeSequence = cms.Sequence( hltL1IsoR9shape + hltL1NonIsoR9shape ) hltMuTrackJpsiPixelTrackSelector.MinMasses = cms.vdouble ( 2.0, 60.0 ) hltMuTrackJpsiPixelTrackSelector.MaxMasses = cms.vdouble ( 4.6, 120.0 ) hltMu5Track1JpsiPixelMassFiltered.MinMasses = cms.vdouble ( 2.0, 60.0 ) hltMu5Track1JpsiPixelMassFiltered.MaxMasses = cms.vdouble ( 4.6, 120.0 ) hltMu5TkMuJpsiTrackMassFiltered.MinMasses = cms.vdouble ( 2.5, 60.0 ) hltMu5TkMuJpsiTrackMassFiltered.MaxMasses = cms.vdouble ( 4.1, 120.0 ) hltMu5Track2JpsiTrackMassFiltered.MinMasses = cms.vdouble ( 2.7, 60.0 ) hltMu5Track2JpsiTrackMassFiltered.MaxMasses = cms.vdouble ( 3.5, 120.0 ) hltMu5L2Mu2JpsiTrackMassFiltered.MinMasses = cms.vdouble ( 1.8, 50.0 ) hltMu5L2Mu2JpsiTrackMassFiltered.MaxMasses = cms.vdouble ( 4.5, 130.0 ) #hltLowMassDisplacedL3Filtered.MaxEta = cms.double(3.0) #hltLowMassDisplacedL3Filtered.MinPtPair = cms.double( 0.0 ) #hltLowMassDisplacedL3Filtered.MinPtMin = cms.double( 0.0 ) #hltLowMassDisplacedL3Filtered.MaxInvMass = cms.double( 11.5 ) # #hltDisplacedmumuFilterLowMass.MinLxySignificance = cms.double( 0.0 ) #hltDisplacedmumuFilterLowMass.MinVtxProbability = cms.double( 0.0 ) #hltDisplacedmumuFilterLowMass.MinCosinePointingAngle = cms.double( -2.0 ) #HLTDisplacemumuSequence = cms.Sequence( hltL1sL1DoubleMu0 + hltDimuonL1Filtered0 + hltDimuonL2PreFiltered0 + hltLowMassDisplacedL3Filtered + hltDisplacedmumuVtxProducerLowMass + hltDisplacedmumuFilterLowMass) # create the jetMET HLT reco path DoHLTJets = cms.Path(HLTBeginSequence + HLTBeginSequence + HLTRecoJetSequenceAK5Corrected + HLTRecoJetSequenceAK5L1FastJetCorrected + HLTRecoMETSequence + HLTDoLocalHcalWithoutHOSequence ) DoHLTJetsU = cms.Path(HLTBeginSequence + HLTBeginSequence + HLTRecoJetSequenceAK5Uncorrected + HLTRecoMETSequence ) # create the muon HLT reco path DoHltMuon = cms.Path( HLTBeginSequence + HLTL2muonrecoSequenceNoVtx + HLTL2muonrecoSequence + HLTL2muonisorecoSequence + HLTL3muonrecoSequence + HLTL3muonisorecoSequence + HLTL3muonTkIso10recoSequence + HLTMuTrackJpsiPixelRecoSequence + HLTMuTrackJpsiTrackRecoSequence + ## HLTDisplacemumuSequence + HLTDoLocalPixelSequence + hltPixelTracks + HLTDoLocalStripSequence + hltMuTrackSeeds + hltMuCkfTrackCandidates + hltMuCtfTracks + hltDiMuonMerging + HLTL3muonrecoNocandSequence + hltDiMuonLinks + hltGlbTrkMuons + hltGlbTrkMuonCands + HLTEndSequence ) # create the Egamma HLT reco paths DoHLTPhoton = cms.Path( HLTBeginSequence + HLTDoRegionalEgammaEcalSequence + HLTL1IsolatedEcalClustersSequence + HLTL1NonIsolatedEcalClustersSequence + hltL1IsoRecoEcalCandidate + hltL1NonIsoRecoEcalCandidate + HLTEgammaR9ShapeSequence + HLTEgammaR9IDSequence + hltL1IsolatedPhotonEcalIsol + hltL1NonIsolatedPhotonEcalIsol + hltL1IsolatedPhotonHcalIsol + hltL1NonIsolatedPhotonHcalIsol + HLTDoLocalPixelSequence + HLTDoLocalStripSequence + hltL1IsoEgammaRegionalPixelSeedGenerator + hltL1IsoEgammaRegionalCkfTrackCandidates + hltL1IsoEgammaRegionalCTFFinalFitWithMaterial + hltL1NonIsoEgammaRegionalPixelSeedGenerator + hltL1NonIsoEgammaRegionalCkfTrackCandidates + hltL1NonIsoEgammaRegionalCTFFinalFitWithMaterial + hltL1IsolatedPhotonHollowTrackIsol + hltL1NonIsolatedPhotonHollowTrackIsol + HLTEcalActivitySequence + hltActivityPhotonHcalForHE + hltActivityR9ID + hltActivityPhotonClusterShape + hltActivityPhotonEcalIsol + hltActivityPhotonHcalIsol + HLTEcalActivityEgammaRegionalRecoTrackerSequence + hltEcalActivityEgammaRegionalAnalyticalTrackSelector + hltActivityPhotonHollowTrackIsolWithId ## hltActivityPhotonHollowTrackIsol ) DoHLTElectron = cms.Path( HLTBeginSequence + HLTDoRegionalEgammaEcalSequence + HLTL1IsolatedEcalClustersSequence + HLTL1NonIsolatedEcalClustersSequence + hltL1IsoRecoEcalCandidate + hltL1NonIsoRecoEcalCandidate + HLTEgammaR9ShapeSequence +#was commented out for HT jobs HLTEgammaR9IDSequence + hltL1IsoHLTClusterShape + hltL1NonIsoHLTClusterShape + hltL1IsolatedPhotonEcalIsol + hltL1NonIsolatedPhotonEcalIsol + HLTDoLocalHcalWithoutHOSequence + hltL1IsolatedPhotonHcalForHE + hltL1NonIsolatedPhotonHcalForHE + hltL1IsolatedPhotonHcalIsol + hltL1NonIsolatedPhotonHcalIsol + HLTDoLocalPixelSequence + HLTDoLocalStripSequence + hltL1IsoStartUpElectronPixelSeeds + hltL1NonIsoStartUpElectronPixelSeeds + hltCkfL1IsoTrackCandidates + hltCtfL1IsoWithMaterialTracks + hltPixelMatchElectronsL1Iso + hltCkfL1NonIsoTrackCandidates + hltCtfL1NonIsoWithMaterialTracks + hltPixelMatchElectronsL1NonIso + hltElectronL1IsoDetaDphi + hltElectronL1NonIsoDetaDphi + HLTL1IsoEgammaRegionalRecoTrackerSequence + HLTL1NonIsoEgammaRegionalRecoTrackerSequence + hltL1IsoElectronTrackIsol + hltL1NonIsoElectronTrackIsol + hltHFEMClusters + hltHFRecoEcalTightCandidate ) # create the tau HLT reco path from HLTrigger.HLTanalyzers.OpenHLT_Tau_cff import * DoHLTTau = cms.Path(HLTBeginSequence + OpenHLTCaloTausCreatorSequence + openhltL2TauJets + openhltL2TauIsolationProducer + HLTDoLocalPixelSequence + HLTRecopixelvertexingSequence + OpenHLTL25TauTrackReconstructionSequence + OpenHLTL25TauTrackIsolation + TauOpenHLT+ HLTRecoJetSequencePrePF + HLTPFJetTriggerSequence + HLTPFJetTriggerSequenceForTaus + HLTPFTauSequence + HLTEndSequence) # create the b-jet HLT paths from HLTrigger.HLTanalyzers.OpenHLT_BJet_cff import * DoHLTBTag = cms.Path( HLTBeginSequence + OpenHLTBLifetimeL25recoSequence + OpenHLTBSoftMuonL25recoSequence + OpenHLTBLifetimeL3recoSequence + OpenHLTBSoftMuonL3recoSequence + HLTEndSequence ) DoHLTAlCaPi0Eta1E31 = cms.Path( HLTBeginSequence + hltL1sAlCaEcalPi0Eta + HLTDoRegionalPi0EtaSequence + HLTEndSequence ) DoHLTAlCaPi0Eta8E29 = cms.Path( HLTBeginSequence + hltL1sAlCaEcalPi0Eta + HLTDoRegionalPi0EtaSequence + HLTEndSequence ) DoHLTAlCaECALPhiSym = cms.Path( HLTBeginSequence + hltEcalRawToRecHitFacility + hltESRawToRecHitFacility + hltEcalRegionalRestFEDs + hltEcalRecHitAll + HLTEndSequence ) DoHLTMinBiasPixelTracks = cms.Path( HLTBeginSequence + HLTDoLocalPixelSequence + HLTDoHILocalPixelSequence + HLTPixelTrackingForHITrackTrigger + hltPixelCandsForHITrackTrigger + hltPixelTracks + hltPixelVertices)
UTF-8
Python
false
false
8,066
py
122
HLTopen_cff.py
75
0.72849
0.704066
0
215
36.511628
210
impiyush83/slambook
11,699,490,929,656
9e6b08fe369787f6b5ffcef866d2389f8212c523
bba538f0246c01407b16fcf990df5e2af0d93cfc
/slambook_restful/resource/customer_homepage.py
683292c7246b425b6916470b5c2d4c82d3aff13c
[]
no_license
https://github.com/impiyush83/slambook
7218f6ed9e3905dcf8531744984c58bba8a9827f
1c3cbb9a6e69199a3978dd51734dd617d1200f4e
refs/heads/master
2022-12-10T23:49:40.962475
2019-07-07T13:50:53
2019-07-07T13:50:53
142,017,369
0
0
null
false
2019-10-24T19:25:28
2018-07-23T13:21:46
2019-10-24T19:23:24
2019-10-24T19:25:27
10,014
0
0
0
CSS
false
false
from flask import current_app as app, render_template, make_response from flask_jwt_extended import jwt_required, get_jwt_identity from flask_restful import Resource from constants.common_constants import headers from slambook_restful.models.Friend.friend import Friend from slambook_restful.models.Secret.secret import Secret from slambook_restful.models.User.user import User from slambook_restful.utils.resource_exceptions import exception_handle from slambook_restful.views.secret_key import get_secret_key class CustomerHomepage(Resource): decorators = [jwt_required, exception_handle] def __init__(self): app.logger.info('In the constructor of {}'.format(self.__class__.__name__)) def get(self): """ .. http:get:: /user/home This api will be used to render homepage **Example request**: .. sourcecode:: http **Example response**: .. sourcecode:: http HTTP/1.1 200 OK Vary: Accept Content-Type: application/json { } :statuscode 200: json :statuscode 400: value error """ user_id = get_jwt_identity() user = User.with_id(user_id) friends = Friend.get_friends_with_email_address(user.id) is_secret_key_created = Secret.is_secret_key_created(user.email) secret_key = get_secret_key(is_secret_key_created) return make_response( render_template('homepage.html', friends=friends, is_secret_key=is_secret_key_created, len=len(friends), secret_key=secret_key), headers)
UTF-8
Python
false
false
1,638
py
30
customer_homepage.py
23
0.647741
0.641026
0
57
27.736842
98
loum/pyrob
15,015,205,680,592
07693ac6d4e671c7fa59e508295a03382acaea1c
fbe55b7d87d11faa7911a23cc78df02bc6af5406
/pyrob/schema/ipact_tran/zabbix/zbx_device_energy.py
fe9e29f2ce9dac378bef77988a2f639fdacc1768
[]
no_license
https://github.com/loum/pyrob
463e25fba2fa6ee1f6da8f256fd42d09a7fc88bd
551ebdfc449fb565d7f324297887401ec553f6c5
refs/heads/master
2020-07-31T12:08:30.095751
2019-09-24T12:34:39
2019-09-24T12:34:39
210,598,588
1
0
null
null
null
null
null
null
null
null
null
null
null
null
null
"""IPACT_TRAN T_TRN_ZBX_DEVICE_ENERGY table class mapping. """ from sqlalchemy import Column, String import sqlalchemy.types import pyrob.table import pyrob.schema.types class ZbxDeviceEnergy(pyrob.table.Common, pyrob.table.Base): __tablename__ = 'T_TRN_ZBX_DEVICE_ENERGY' __table_args__ = {'schema': 'IPACT_TRAN'} physical_name = Column(String(25), primary_key=True, nullable=False) energy_consumed = Column(pyrob.schema.types.Float) battery_temperature = Column(pyrob.schema.types.Float) mains_voltage = Column(pyrob.schema.types.Float) time_captured_utc = Column(pyrob.schema.types.OraDateParser, primary_key=True, nullable=False) battery_current = Column(pyrob.schema.types.Float) load_current = Column(pyrob.schema.types.Float) output_voltage = Column(pyrob.schema.types.Float) rectifier_current = Column(pyrob.schema.types.Float) ipact_dl_process_id = Column(pyrob.schema.types.Int) record_created_userid = Column(pyrob.schema.types.Int) record_created_date = Column(pyrob.schema.types.OraDateParser) @sqlalchemy.ext.hybrid.hybrid_property def source_table_details(self): return { self.__tablename__: ( 'PHYSICAL_NAME', 'TIME_CAPTURED_UTC', 'ENERGY_CONSUMED', ) } @sqlalchemy.ext.hybrid.hybrid_property def source_table_order_by(self): return ( 'PHYSICAL_NAME', 'TIME_CAPTURED_UTC', ) @sqlalchemy.ext.hybrid.hybrid_property def source_table_date_range_col(self): return 'TIME_CAPTURED_UTC' @sqlalchemy.event.listens_for(ZbxDeviceEnergy, 'before_insert') def gen_default(mapper, connection, instance): if instance.energy_consumed is None: instance.energy_consumed = instance.output_voltage * instance.rectifier_current * 0.25
UTF-8
Python
false
false
1,931
py
682
zbx_device_energy.py
487
0.659762
0.657172
0
55
34.109091
94
Naught0/qtbot
14,001,593,405,335
075c185debf3d3aa830895906adaa860d768ab95
5616e441a3bd34cb46ab211ce625609b4cebca3c
/cogs/news.py
dcbc1b8f7c718352870054505a6b8b4e54793ed6
[ "MIT" ]
permissive
https://github.com/Naught0/qtbot
08a55ae758f8f42611cdc186fc37749ec9d3e646
d5824aebc4892b3fbe24c2d053dbb4052c57b5ae
refs/heads/master
2023-06-24T15:51:22.505671
2023-06-15T19:19:41
2023-06-15T19:45:12
93,383,576
9
9
MIT
false
2023-05-23T00:28:18
2017-06-05T08:43:16
2023-02-20T17:32:04
2023-05-23T00:28:18
5,100
3
1
2
Python
false
false
import json import asyncio import discord from urllib.parse import quote from dateutil.parser import parse from datetime import datetime, timedelta from utils import aiohttp_wrap as aw from discord.ext import commands class News(commands.Cog): def __init__(self, bot): self.bot = bot self.redis_client = bot.redis_client self.aio_session = bot.aio_session self.uri = "http://api.mediastack.com/v1/news" with open("data/apikeys.json") as f: self.api_key = json.load(f)["news"] @staticmethod def date_range() -> str: return f"{(datetime.today() - timedelta(days=5)).isoformat().split('T')[0]},{datetime.today().isoformat().split('T')[0]}" @staticmethod def json_to_embed(json_dict: dict) -> discord.Embed: em = discord.Embed() em.title = json_dict["title"] em.description = json_dict["description"] em.url = json_dict["url"] # This field is empty sometimes -> handle it if json_dict["image"]: em.set_thumbnail(url=json_dict["image"]) em.set_footer(text=json_dict["source"]) em.timestamp = parse(json_dict["published_at"]) return em @commands.command(name="news") async def get_news( self, ctx: commands.Context, *, query: commands.clean_content(escape_markdown=True) = None, ): """Get the latest and greatest news or optionally search for some specific news stories""" # Add Emojis for navigation emoji_tup = tuple(f"{x}\U000020e3" for x in range(1, 10)) em_dict = {} params = ( { "keywords": quote(query), "languages": "en", "limit": 9, "date": self.date_range(), "sort": "popularity", "categories": "-sports,-entertainment,-business", "access_key": self.api_key, } if query else { "languages": "en", "limit": 9, "date": self.date_range(), "sort": "popularity", "categories": "-sports,-entertainment,-business", "countries": "us,gb", "access_key": self.api_key, } ) redis_key = f"news:{query}" if query else "news" if await self.redis_client.exists(redis_key): raw_json_string = await self.redis_client.get(redis_key) raw_json_dict = json.loads(raw_json_string) article_list = raw_json_dict["data"] for idx, article in enumerate(article_list[:9]): em_dict[emoji_tup[idx]] = self.json_to_embed(article) else: api_response = await aw.aio_get_json( self.aio_session, self.uri, params=params, ) if api_response is None: return await ctx.error( "API error", description="Something went wrong with that request. Try again later.", ) article_list = api_response["data"] if len(article_list) == 0: return await ctx.error( "No articles found", description=f"Couldn't find any news on `{query}`", ) await self.redis_client.set(redis_key, json.dumps(api_response), ex=10 * 60) for idx, article in enumerate(article_list): em_dict[emoji_tup[idx]] = self.json_to_embed(article) bot_message = await ctx.send(embed=em_dict[emoji_tup[0]]) for emoji in emoji_tup[: len(article_list)]: await bot_message.add_reaction(emoji) def check(reaction, user): return user == ctx.author and reaction.emoji in emoji_tup and reaction.message.id == bot_message.id while True: try: reaction, user = await self.bot.wait_for("reaction_add", check=check, timeout=30.0) except asyncio.TimeoutError: return await bot_message.clear_reactions() if reaction.emoji in em_dict: await bot_message.edit(embed=em_dict[reaction.emoji]) await bot_message.remove_reaction(reaction.emoji, ctx.author) def setup(bot): bot.add_cog(News(bot))
UTF-8
Python
false
false
4,396
py
59
news.py
51
0.544586
0.538672
0
131
32.557252
129
fcea/scrape_InflFore_Lid
4,604,204,973,838
7855838359b0713de7ad99f8485eb617a5f8f706
5954f1181b3654cbfdaea011326013af769ea012
/Lider_Webscraper_Alimentos.py
f6c3aa3ec4eec743e5baa3a22efd69bd5884d51a
[]
no_license
https://github.com/fcea/scrape_InflFore_Lid
7db2873ed9256db5c268f6d5acb0f35438d9c0bd
1e8dc8f3a976d485d73bb7a2172b5ed1a04e927b
refs/heads/master
2021-01-22T20:43:54.928600
2013-08-28T05:46:30
2013-08-28T05:46:30
11,716,225
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
import csv from math import ceil import datetime import re from selenium import webdriver from selenium.webdriver.common.keys import Keys import datetime from bs4 import UnicodeDammit import MySQLdb as mdb import time import sys import random #METODO QUE CREA LA TABLA SQL EN CASO DE QUE NO EXISTA def create_Table_Productos(name): con = mdb.connect('localhost', 'root', 'password', 'LiderWebscraping'); with con: cur = con.cursor() #cur.execute("DROP TABLE IF EXISTS "+tableName) cur.execute("CREATE TABLE IF NOT EXISTS "+name+"(SKU VARCHAR(50),\ Categoria VARCHAR(200) ,\ Cod_Categoria INT ,\ Marca VARCHAR(200) ,\ Producto VARCHAR(1000) ,\ FechaEjecucion VARCHAR(20) ,\ Disponibilidad INT ,\ PrecioProducto INT ,\ PrecioGranel INT ,\ PrecioNormal INT)") con.close() #METODO QUE COORDINA EL SCRAPING DE TODO EL SITIO WEB #RECORRE LOS LINKS, DETERMINA LOS QUE SIRVEN. #DE LOS QUE SIRVEN VERIFICA CUALES AUN NO HAN SIDO VISITADOS #HACE LA REVISION Y DESPUES LLAMA A LA FUNCION QUE INSERTA EN LA TABLA def lets_Scrape_Supermarket(url, fecha_ejec, tableName): driver = webdriver.Firefox() driver.get(url) superRef = driver.find_element_by_xpath('/html/body/div[3]/div[3]/ul/li/a') superRef.click() links = driver.find_elements_by_css_selector(".cursorpointer.alertCheck") linksTmp =[] nombresTmp = [] cont=0 for ref in links: temp = ref.get_attribute("outerHTML").encode('utf-8').strip() #BGood y Experto en la Parrilla ya fueron guardados y corresponden a vistas especiales que lider genera para reordenar productos existentes #"Huggies" y "Vinos y Mas" son categorias no visitadas pero que tienen una vista distinta. En este caso entonces se procede a generar / inputar el link directamente. if ref.get_attribute("innerHTML").encode('utf-8').strip() != "B Good" and ref.get_attribute("innerHTML").encode('utf-8').strip() != "Experto en la Parrilla" and temp.find("categoryFood") == -1 and temp.find("specialCategoryFood") == -1 and temp.find("CAT_GM") == -1: if ref.get_attribute("innerHTML").encode('utf-8').strip() == "Huggies": linksTmp.append('http://www.lider.cl/dys/catalog/product/productList.jsp?id=cat870020') elif ref.get_attribute("innerHTML").encode('utf-8').strip()[:9] == "Vinos y M": linksTmp.append('http://www.lider.cl/dys/catalog/product/productList.jsp?id=CF_Nivel3_000159') else: linksTmp.append("http://www.lider.cl" + get_Link(temp)) nombresTmp.append(str(cont) + "|" + ref.get_attribute("innerHTML").encode('utf-8').strip()) cont = cont+1 #driver.close() cont = 0 #busco id categoria del link mas alta ingresada, asigno el valor al contador e itero desde ese punto cont = data_Inserted(tableName) while cont < len(linksTmp): print "Ejecuto es " + nombresTmp[cont] resuPart = get_info_Supermarket(driver, linksTmp[cont], fecha_ejec, nombresTmp[cont], tableName) #recupero 7 elementos a insertar, falta la fecha, categoria y cod_cate, pero estan ya aca cont = cont + 1 driver.close() #METODO QUE SIRVE PARA CAMBIAR EL FORMATO DE LOS NUMEROS QUE RECUPERO DESDE LIDER #RETORNA LA FECHA SIN SIGNOS $ NI PUNTOS def modify_Prices(formPrice): form1 = formPrice.replace(".","") form2 = form1.replace("$","") return form2 #METODO QUE SIRVE PARA VERIFICAR EL ULTIMO DATO INSERTADO EN LA TABLA SQL #RETORNA 0 SI LA TABLA ESTA VACIA. DE LO CONTRARIO RETORNA EL CODIGO MAS ALTO INGRESADO. def data_Inserted(tableName): con = mdb.connect('localhost', 'root', 'password', 'LiderWebscraping'); with con: cur = con.cursor() cur.execute("SELECT Cod_Categoria FROM " +tableName) #Esto me hace pasar la lista a un arreglo row = [item[0] for item in cur.fetchall()] con.close() try: maxim = max(row) return maxim except ValueError: return 0 #METODO AUXILIAR PARA DEJAR EL LINK CON EL FORMATO APROPIADO #RETORNA EL LINK CON TRIMS def get_Link(unfText): st = unfText.find('lang')+6 nd = unfText.find('"', unfText.find('lang')+6) return unfText[st:nd] #METODO QUE HACE EL INSERT DENTRO DE LA TABLA SQL def word_to_SQL_insertion(concatString, tableName): con = mdb.connect('localhost', 'root', 'password', 'LiderWebscraping'); with con: cur = con.cursor() if cur.execute("SELECT * FROM "+tableName+" WHERE SKU=%s",concatString.split("|")[0]) == 0: cur.execute("INSERT INTO "+tableName+" VALUES (%s,%s,%s,%s, %s,%s,%s,%s,%s, %s)",(concatString.split("|")[0],concatString.split("|")[1],int(concatString.split("|")[2]), concatString.split("|")[3],concatString.split("|")[4],concatString.split("|")[5],int(concatString.split("|")[6]),int(concatString.split("|")[7]),int(concatString.split("|")[8]),int(concatString.split("|")[9]))) else: #print 'Nada que hacer con ' + str(concatString.split("|")[0]) pass con.close() def get_Seven_Data(ele1): #Obtengo el SKU. No hay metodo eficiente, solo busqueda en string auxText = ele1.get_attribute("outerHTML").encode('utf-8').strip() skuProd = auxText[auxText.find("skuId=")+6:auxText.find("&",auxText.find("skuId=")) ] #Saco ahora la informacion de los precios y otros. Aprovecho la estructura del texto. info = ele1.find_elements_by_tag_name("a") #info[0] -> data inutil #info[1] -> marca #info[2] -> detalle #info[3] -> carro de compras #info[4] -> carro de compras #info[5] -> solo los productos que estan disponibles #De este modo, obtengo la marca y el detalle try: marcaDirt = info[1].get_attribute("innerHTML").encode('utf-8').strip() exist =1 except: exist = 0 if exist == 1: marcaProd = marcaDirt[marcaDirt.find(">")+1:marcaDirt.find("<",marcaDirt.find(">"))] detalleProd = info[2].get_attribute("innerHTML").encode('utf-8').strip() #La existencia de la clase bloqueada me indica si el producto esta actualmente disponible. try: availInd = ele1.find_element_by_class_name("ech_form_disabledDiv").get_attribute("innerHTML").encode('utf-8').strip() availInd = 0 except: availInd = 1 try: tiposPrecios = ele1.find_elements_by_class_name("retail") if len(tiposPrecios)==0: #significa que hay solo precio primario mainPrice = modify_Prices(ele1.find_element_by_class_name("price").get_attribute("innerHTML").encode('utf-8').strip()) granelPrice = modify_Prices(ele1.find_element_by_class_name("price").get_attribute("innerHTML").encode('utf-8').strip()) normalPrice = modify_Prices(ele1.find_element_by_class_name("price").get_attribute("innerHTML").encode('utf-8').strip()) elif len(tiposPrecios)==1: #significa que hay precio primario y precio granel auxPrice=tiposPrecios[0].get_attribute("innerHTML").encode('utf-8').strip() mainPrice = modify_Prices(ele1.find_element_by_class_name("price").get_attribute("innerHTML").encode('utf-8').strip()) granelPrice = modify_Prices(re.findall('\${1}\s*[\.0-9]{1,10}\s*', auxPrice)[0]).strip() normalPrice = modify_Prices(re.findall('\${1}\s*[\.0-9]{1,10}\s*', auxPrice)[0]).strip() elif len(tiposPrecios)==2: #significa que hay precio primario, granel y normal (ergo el primario esta en oferta) #el precio no lo puedo sacar directamente, tengo que hacer un parser manual sobre el texto outer completo hiddenPrice = ele1.get_attribute("outerHTML").encode('utf-8').strip().split('<small class="retail">') maskedPrice = tiposPrecios[1].get_attribute("outerHTML").encode('utf-8').strip() mainPrice = modify_Prices(ele1.find_element_by_class_name("price").get_attribute("innerHTML").encode('utf-8').strip()) granelPrice = modify_Prices(re.findall('\${1}\s*[\.0-9]{1,10}\s*', maskedPrice)[0]).strip() normalPrice = modify_Prices(hiddenPrice[1][hiddenPrice[1].find('$'):].strip()) else: #caso no determinado print 'Hay un caso que no se esta capturando!' sys.exit(0) except: #significa que no hay ningun tag con ese nombre mainPrice = modify_Prices(ele1.find_element_by_class_name("price").get_attribute("innerHTML").encode('utf-8').strip()) granelPrice = modify_Prices(ele1.find_element_by_class_name("price").get_attribute("innerHTML").encode('utf-8').strip()) normalPrice = modify_Prices(ele1.find_element_by_class_name("price").get_attribute("innerHTML").encode('utf-8').strip()) return [skuProd , marcaProd, detalleProd, availInd, mainPrice, granelPrice, normalPrice] else: return [0] #METODO QUE RECUPERA LA INFORMACION DESDE LA PAGINA (HACE EL WEBSCRAPE ) #RETORNA 7 PARAMETROS DE INFORMACION AL METODO PRINCIPAL. VAN SEPARADOS POR "|" PARA SPLITEAR DESPUES def get_info_Supermarket(driver, urls, fecha_ejec, doubleParam, tableName): #genero la url correcta con todos los elementos cargados #driver = webdriver.Firefox() driver.get(urls) longit = driver.find_element_by_class_name('pages-shopping').find_element_by_class_name('info') numUnclean = longit.get_attribute("innerHTML").encode('utf-8').strip().split('<strong>')[2] numElem = numUnclean[:numUnclean.find('<')] urrs = urls[:urls.find('&')]+ '&pageSize=' + str(numElem) + '&goToPage=1' #tengo la url nueva correcta driver.get(urrs) #Recupero informacion element2 = driver.find_elements_by_class_name('prod_referencia') #sku element3 = driver.find_elements_by_class_name('price') #precio rojo #captura la informacion de todos los productos no featured elementemp = driver.find_elements_by_css_selector('.product.ech_prod') #recupero la marca, detalle y si tiene stock. #captura la informacion de todos los productos featured elementfeat = driver.find_elements_by_css_selector('.product.first.prominent_block.no-borde') #recupero la marca, detalle y si tiene stock. #Para productos featured for ele1 in elementfeat: retorno = get_Seven_Data(ele1) #Es 0 solo en el caso que haya grilla pero no producto if retorno[0] != 0: #Proceso los parametros que no tengo codeCategory = doubleParam.split("|")[0] nameCategory = doubleParam.split("|")[1] #Hago el llamado a la funcion que inserta en la base de datos inputSQL =retorno[0] +"|"+ nameCategory +"|"+ str(codeCategory) +"|"+ retorno[1] +"|"+ retorno[2] +"|"+ fecha_ejec +"|"+ str(retorno[3]) +"|"+ str(retorno[4]) +"|"+ str(retorno[5]) +"|"+ str(retorno[6]) word_to_SQL_insertion(inputSQL, tableName) else: pass #Para productos no featured for ele1 in elementemp: retorno = get_Seven_Data(ele1) #Es 0 solo en el caso que haya grilla pero no producto if retorno[0] != 0: #Proceso los parametros que no tengo codeCategory = doubleParam.split("|")[0] nameCategory = doubleParam.split("|")[1] #Hago el llamado a la funcion que inserta en la base de datos inputSQL =retorno[0] +"|"+ nameCategory +"|"+ str(codeCategory) +"|"+ retorno[1] +"|"+ retorno[2] +"|"+ fecha_ejec +"|"+ str(retorno[3]) +"|"+ str(retorno[4]) +"|"+ str(retorno[5]) +"|"+ str(retorno[6]) word_to_SQL_insertion(inputSQL, tableName) else: pass #driver.close() return #MAIN if __name__ == '__main__': #fecha del dia d = datetime.datetime.now().strftime("%Y-%m-%d") d_name = d.replace("-","") #inicializo tabla en MySQL LiderSupermarket = d_name + '_Lider_Supermarket' create_Table_Productos(LiderSupermarket) #ejecuto el proceso lets_Scrape_Supermarket('http://www.lider.cl/dys/',d,LiderSupermarket)
UTF-8
Python
false
false
11,249
py
8
Lider_Webscraper_Alimentos.py
7
0.698551
0.682994
0
235
46.868085
386
kishmakov/rs
1,425,929,181,365
25fdd4766fe77277721e608e413b18eaa25ac1e1
1d62731ed2ebc3c84ed27d40c7d0417f6ce0ab0e
/rs/urls.py
b484a9e53dfeca1d32c3c23fc53ab0c39b628994
[]
no_license
https://github.com/kishmakov/rs
064fd19aeac4dafda8788a35aa3bb5c5e78c687c
c5e932c36fd2a62625aaf5b374f754ee8b839602
refs/heads/master
2021-01-10T08:25:21.901462
2013-03-23T18:04:55
2013-03-23T18:04:55
null
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
from django.conf.urls import patterns from app.views import process_beg, process_go import settings # Uncomment the next two lines to enable the admin: # from django.contrib import admin # admin.autodiscover() urlpatterns = patterns('', ('^$', process_beg), ('^go/$', process_go), (r'^static/(?P<path>.*)$', 'django.views.static.serve', {'document_root': settings.STATIC_ROOT}), )
UTF-8
Python
false
false
392
py
21
urls.py
4
0.693878
0.693878
0
13
29.153846
100
eco3/stochastik-ss21
7,662,221,700,551
3c34e698f52145225364495abc042a20ce008a31
7b43998db0d31e0e2036f4c06681d263a7875f5b
/Klausurtools/tools/multivariate_statistik.py
efbeeb0cd43a2cdd07aaa60c0c5c5bcab596747e
[]
no_license
https://github.com/eco3/stochastik-ss21
3dbabf87495a6274730b799c630a3b98624766cc
3974a51c3b7c5dc9ebc79df01fc14cf635da9b11
refs/heads/master
2023-06-16T23:46:31.069291
2021-07-14T16:25:50
2021-07-14T16:25:50
353,711,310
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
import numpy as np from scipy import stats import matplotlib.pyplot as plt def lineare_korrelation(a, b): korrelationskoeffizient = np.corrcoef(a, b)[0][1] kovarianz = np.cov(a, b)[0][1] print(f"(empirischer) Korrelationskoeffizient\t", "np.corrcoef(...)\t", f"r_x,y = {korrelationskoeffizient}") print(f"(empirische) Kovarianz\t\t\t", "np.cov(...)\t\t", f"s_x,y = {kovarianz}") def lineare_regression(a, b): reg = stats.linregress(a, b) print("Lineare Regression") print(f"f(x) = {reg.slope:.3f} * x + {reg.intercept:.3f}") print(f"Bestimmheitsmaß: R² = {reg.rvalue**2}") def plot_regression(a, b): reg = stats.linregress(a, b) plt.plot(a, b, 'o', label='original data') plt.plot(a, reg.intercept + reg.slope * np.asarray(a), 'r', label='fitted line') plt.legend() plt.show()
UTF-8
Python
false
false
840
py
13
multivariate_statistik.py
11
0.636038
0.626492
0
28
28.928571
113
AtomikKeo/ProgrammingContestPractice
3,332,894,629,908
8d539bbd68cf165857879ea93afc6144826e74ab
e0c5d703ab688e6a1f8384279d9ce46cf0f667c7
/AtCoder/ABC/66/ss.py
733f483d321037ca63be9df18d69be72e76ad7da
[]
no_license
https://github.com/AtomikKeo/ProgrammingContestPractice
a9b739156981f8c1b7213549ff2f258c4c6af8d4
52140087e825fad3be3a19c73814e70c434e0f0c
refs/heads/master
2018-04-05T04:57:07.735681
2018-02-26T14:27:21
2018-02-26T14:27:21
89,137,597
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
s = input() for i in range(1, len(s)): if((len(s)-i) % 2 == 0): ans = len(s) - i if(s[0:ans//2] == s[ans//2:ans]): print(ans) break
UTF-8
Python
false
false
176
py
66
ss.py
66
0.386364
0.352273
0
7
24.142857
41
bbmhmmad/django
6,820,408,105,748
652069c451fc365fd1b0c38b582906b43b44c447
f3100ddd6c7f20538949785fce97960ba07f4128
/poke_app/apps/poke_app/migrations/0001_initial.py
be7be1cc62107330aaf1baa49dd56cf8dbba01db
[]
no_license
https://github.com/bbmhmmad/django
bfb14ea307515ae8675d404195dae9374e1a8550
5b6cc59bb2c40d7f4c7fc57e8b02e0c8c477ff9c
refs/heads/master
2021-01-17T11:20:07.160572
2017-03-07T08:58:51
2017-03-07T08:58:51
84,030,826
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
# -*- coding: utf-8 -*- # Generated by Django 1.10.3 on 2016-11-18 16:01 from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): initial = True dependencies = [ ('login_app', '0001_initial'), ] operations = [ migrations.CreateModel( name='Poke', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('created_at', models.DateField(auto_now_add=True)), ('updated_at', models.DateField(auto_now=True)), ('poked', models.ManyToManyField(related_name='user_poked', to='login_app.User')), ('poker', models.ManyToManyField(related_name='user_poker', to='login_app.User')), ], ), ]
UTF-8
Python
false
false
860
py
23
0001_initial.py
12
0.576744
0.552326
0
27
30.851852
114
ash9750/Re0-ML
11,312,943,867,671
db98133ed3d5d503bc92008b237a826b6fcb41f7
bf1f5acf58438c0eeed4e822ff5dfcafa4d5f934
/balanced_kdtree.py
80f1aa23d128ea883e05caebb1acf2cd8300c3aa
[]
no_license
https://github.com/ash9750/Re0-ML
1d5ac183a724461ddd81af5bbc4db5ae482c33c4
094cc72b16bf55892213ec355de4199033d4ff06
refs/heads/master
2021-01-21T11:15:12.079280
2019-04-06T13:26:19
2019-04-06T13:26:19
83,539,185
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
class Node: def __init__(self, point, dimension, depth, left=[], right=[]): self.left = left self.right = right self.point = point self.dimension = dimension self.depth = depth def median(data): m = int(len(data) / 2) return data[m], m def build_tree(data, dimension, node_list, k, depth=0): if depth < k: data = sorted(data, key=lambda x: x[dimension]) data_median, index_medina = median(data) del data[index_medina] node_list.append(Node(data_median, dimension, depth, left=data[:index_medina], right=data[index_medina:])) depth += 1 if index_medina > 0: build_tree(data[:index_medina], not dimension, node_list, k, depth=depth) if len(data) > 1: build_tree(data[index_medina:], not dimension, node_list, k, depth=depth) return node_list def train(data): node_list = [] dimension = False node_list = build_tree(data, dimension, node_list, k=3) for i in range(len(node_list)): print("node:", i, "\tpoint:", node_list[i].point, "\tdepth:", node_list[i].depth, "\tdimension:", node_list[i].dimension, "\tleft:", node_list[i].left, "right:", node_list[i].right) T = [[2, 3], [5, 4], [9, 6], [4, 7], [8, 1], [7, 2]] train(T)
UTF-8
Python
false
false
1,369
py
4
balanced_kdtree.py
3
0.552958
0.53981
0
40
32.225
102
aCoffeeYin/pyreco
16,140,487,119,963
31977ba1fa30d78438c128d5ce598b796198cfc1
9b9a02657812ea0cb47db0ae411196f0e81c5152
/repoData/openembedded-oe-core/allPythonContent.py
a0f23ebc945949a2b8fe7601b1af6dc68986c47f
[]
no_license
https://github.com/aCoffeeYin/pyreco
cb42db94a3a5fc134356c9a2a738a063d0898572
0ac6653219c2701c13c508c5c4fc9bc3437eea06
refs/heads/master
2020-12-14T14:10:05.763693
2016-06-27T05:15:15
2016-06-27T05:15:15
null
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
__FILENAME__ = buildhistory_analysis # Report significant differences in the buildhistory repository since a specific revision # # Copyright (C) 2012 Intel Corporation # Author: Paul Eggleton <paul.eggleton@linux.intel.com> # # Note: requires GitPython 0.3.1+ # # You can use this from the command line by running scripts/buildhistory-diff # import sys import os.path import difflib import git import re import bb.utils # How to display fields list_fields = ['DEPENDS', 'RPROVIDES', 'RDEPENDS', 'RRECOMMENDS', 'RSUGGESTS', 'RREPLACES', 'RCONFLICTS', 'FILES', 'FILELIST', 'USER_CLASSES', 'IMAGE_CLASSES', 'IMAGE_FEATURES', 'IMAGE_LINGUAS', 'IMAGE_INSTALL', 'BAD_RECOMMENDATIONS', 'PACKAGE_EXCLUDE'] list_order_fields = ['PACKAGES'] defaultval_map = {'PKG': 'PKG', 'PKGE': 'PE', 'PKGV': 'PV', 'PKGR': 'PR'} numeric_fields = ['PKGSIZE', 'IMAGESIZE'] # Fields to monitor monitor_fields = ['RPROVIDES', 'RDEPENDS', 'RRECOMMENDS', 'RREPLACES', 'RCONFLICTS', 'PACKAGES', 'FILELIST', 'PKGSIZE', 'IMAGESIZE', 'PKG'] ver_monitor_fields = ['PKGE', 'PKGV', 'PKGR'] # Percentage change to alert for numeric fields monitor_numeric_threshold = 10 # Image files to monitor (note that image-info.txt is handled separately) img_monitor_files = ['installed-package-names.txt', 'files-in-image.txt'] # Related context fields for reporting (note: PE, PV & PR are always reported for monitored package fields) related_fields = {} related_fields['RDEPENDS'] = ['DEPENDS'] related_fields['RRECOMMENDS'] = ['DEPENDS'] related_fields['FILELIST'] = ['FILES'] related_fields['PKGSIZE'] = ['FILELIST'] related_fields['files-in-image.txt'] = ['installed-package-names.txt', 'USER_CLASSES', 'IMAGE_CLASSES', 'ROOTFS_POSTPROCESS_COMMAND', 'IMAGE_POSTPROCESS_COMMAND'] related_fields['installed-package-names.txt'] = ['IMAGE_FEATURES', 'IMAGE_LINGUAS', 'IMAGE_INSTALL', 'BAD_RECOMMENDATIONS', 'NO_RECOMMENDATIONS', 'PACKAGE_EXCLUDE'] class ChangeRecord: def __init__(self, path, fieldname, oldvalue, newvalue, monitored): self.path = path self.fieldname = fieldname self.oldvalue = oldvalue self.newvalue = newvalue self.monitored = monitored self.related = [] self.filechanges = None def __str__(self): return self._str_internal(True) def _str_internal(self, outer): if outer: if '/image-files/' in self.path: prefix = '%s: ' % self.path.split('/image-files/')[0] else: prefix = '%s: ' % self.path else: prefix = '' def pkglist_combine(depver): pkglist = [] for k,v in depver.iteritems(): if v: pkglist.append("%s (%s)" % (k,v)) else: pkglist.append(k) return pkglist if self.fieldname in list_fields or self.fieldname in list_order_fields: if self.fieldname in ['RPROVIDES', 'RDEPENDS', 'RRECOMMENDS', 'RSUGGESTS', 'RREPLACES', 'RCONFLICTS']: (depvera, depverb) = compare_pkg_lists(self.oldvalue, self.newvalue) aitems = pkglist_combine(depvera) bitems = pkglist_combine(depverb) else: aitems = self.oldvalue.split() bitems = self.newvalue.split() removed = list(set(aitems) - set(bitems)) added = list(set(bitems) - set(aitems)) if removed or added: if removed and not bitems: out = '%s: removed all items "%s"' % (self.fieldname, ' '.join(removed)) else: out = '%s:%s%s' % (self.fieldname, ' removed "%s"' % ' '.join(removed) if removed else '', ' added "%s"' % ' '.join(added) if added else '') else: out = '%s changed order' % self.fieldname elif self.fieldname in numeric_fields: aval = int(self.oldvalue or 0) bval = int(self.newvalue or 0) if aval != 0: percentchg = ((bval - aval) / float(aval)) * 100 else: percentchg = 100 out = '%s changed from %s to %s (%s%d%%)' % (self.fieldname, self.oldvalue or "''", self.newvalue or "''", '+' if percentchg > 0 else '', percentchg) elif self.fieldname in defaultval_map: out = '%s changed from %s to %s' % (self.fieldname, self.oldvalue, self.newvalue) if self.fieldname == 'PKG' and '[default]' in self.newvalue: out += ' - may indicate debian renaming failure' elif self.fieldname in ['pkg_preinst', 'pkg_postinst', 'pkg_prerm', 'pkg_postrm']: if self.oldvalue and self.newvalue: out = '%s changed:\n ' % self.fieldname elif self.newvalue: out = '%s added:\n ' % self.fieldname elif self.oldvalue: out = '%s cleared:\n ' % self.fieldname alines = self.oldvalue.splitlines() blines = self.newvalue.splitlines() diff = difflib.unified_diff(alines, blines, self.fieldname, self.fieldname, lineterm='') out += '\n '.join(list(diff)[2:]) out += '\n --' elif self.fieldname in img_monitor_files or '/image-files/' in self.path: fieldname = self.fieldname if '/image-files/' in self.path: fieldname = os.path.join('/' + self.path.split('/image-files/')[1], self.fieldname) out = 'Changes to %s:\n ' % fieldname else: if outer: prefix = 'Changes to %s ' % self.path out = '(%s):\n ' % self.fieldname if self.filechanges: out += '\n '.join(['%s' % i for i in self.filechanges]) else: alines = self.oldvalue.splitlines() blines = self.newvalue.splitlines() diff = difflib.unified_diff(alines, blines, fieldname, fieldname, lineterm='') out += '\n '.join(list(diff)) out += '\n --' else: out = '%s changed from "%s" to "%s"' % (self.fieldname, self.oldvalue, self.newvalue) if self.related: for chg in self.related: if not outer and chg.fieldname in ['PE', 'PV', 'PR']: continue for line in chg._str_internal(False).splitlines(): out += '\n * %s' % line return '%s%s' % (prefix, out) class FileChange: changetype_add = 'A' changetype_remove = 'R' changetype_type = 'T' changetype_perms = 'P' changetype_ownergroup = 'O' changetype_link = 'L' def __init__(self, path, changetype, oldvalue = None, newvalue = None): self.path = path self.changetype = changetype self.oldvalue = oldvalue self.newvalue = newvalue def _ftype_str(self, ftype): if ftype == '-': return 'file' elif ftype == 'd': return 'directory' elif ftype == 'l': return 'symlink' elif ftype == 'c': return 'char device' elif ftype == 'b': return 'block device' elif ftype == 'p': return 'fifo' elif ftype == 's': return 'socket' else: return 'unknown (%s)' % ftype def __str__(self): if self.changetype == self.changetype_add: return '%s was added' % self.path elif self.changetype == self.changetype_remove: return '%s was removed' % self.path elif self.changetype == self.changetype_type: return '%s changed type from %s to %s' % (self.path, self._ftype_str(self.oldvalue), self._ftype_str(self.newvalue)) elif self.changetype == self.changetype_perms: return '%s changed permissions from %s to %s' % (self.path, self.oldvalue, self.newvalue) elif self.changetype == self.changetype_ownergroup: return '%s changed owner/group from %s to %s' % (self.path, self.oldvalue, self.newvalue) elif self.changetype == self.changetype_link: return '%s changed symlink target from %s to %s' % (self.path, self.oldvalue, self.newvalue) else: return '%s changed (unknown)' % self.path def blob_to_dict(blob): alines = blob.data_stream.read().splitlines() adict = {} for line in alines: splitv = [i.strip() for i in line.split('=',1)] if len(splitv) > 1: adict[splitv[0]] = splitv[1] return adict def file_list_to_dict(lines): adict = {} for line in lines: # Leave the last few fields intact so we handle file names containing spaces splitv = line.split(None,4) # Grab the path and remove the leading . path = splitv[4][1:].strip() # Handle symlinks if(' -> ' in path): target = path.split(' -> ')[1] path = path.split(' -> ')[0] adict[path] = splitv[0:3] + [target] else: adict[path] = splitv[0:3] return adict def compare_file_lists(alines, blines): adict = file_list_to_dict(alines) bdict = file_list_to_dict(blines) filechanges = [] for path, splitv in adict.iteritems(): newsplitv = bdict.pop(path, None) if newsplitv: # Check type oldvalue = splitv[0][0] newvalue = newsplitv[0][0] if oldvalue != newvalue: filechanges.append(FileChange(path, FileChange.changetype_type, oldvalue, newvalue)) # Check permissions oldvalue = splitv[0][1:] newvalue = newsplitv[0][1:] if oldvalue != newvalue: filechanges.append(FileChange(path, FileChange.changetype_perms, oldvalue, newvalue)) # Check owner/group oldvalue = '%s/%s' % (splitv[1], splitv[2]) newvalue = '%s/%s' % (newsplitv[1], newsplitv[2]) if oldvalue != newvalue: filechanges.append(FileChange(path, FileChange.changetype_ownergroup, oldvalue, newvalue)) # Check symlink target if newsplitv[0][0] == 'l': if len(splitv) > 3: oldvalue = splitv[3] else: oldvalue = None newvalue = newsplitv[3] if oldvalue != newvalue: filechanges.append(FileChange(path, FileChange.changetype_link, oldvalue, newvalue)) else: filechanges.append(FileChange(path, FileChange.changetype_remove)) # Whatever is left over has been added for path in bdict: filechanges.append(FileChange(path, FileChange.changetype_add)) return filechanges def compare_lists(alines, blines): removed = list(set(alines) - set(blines)) added = list(set(blines) - set(alines)) filechanges = [] for pkg in removed: filechanges.append(FileChange(pkg, FileChange.changetype_remove)) for pkg in added: filechanges.append(FileChange(pkg, FileChange.changetype_add)) return filechanges def compare_pkg_lists(astr, bstr): depvera = bb.utils.explode_dep_versions2(astr) depverb = bb.utils.explode_dep_versions2(bstr) # Strip out changes where the version has increased remove = [] for k in depvera: if k in depverb: dva = depvera[k] dvb = depverb[k] if dva and dvb and len(dva) == len(dvb): # Since length is the same, sort so that prefixes (e.g. >=) will line up dva.sort() dvb.sort() removeit = True for dvai, dvbi in zip(dva, dvb): if dvai != dvbi: aiprefix = dvai.split(' ')[0] biprefix = dvbi.split(' ')[0] if aiprefix == biprefix and aiprefix in ['>=', '=']: if bb.utils.vercmp(bb.utils.split_version(dvai), bb.utils.split_version(dvbi)) > 0: removeit = False break else: removeit = False break if removeit: remove.append(k) for k in remove: depvera.pop(k) depverb.pop(k) return (depvera, depverb) def compare_dict_blobs(path, ablob, bblob, report_all, report_ver): adict = blob_to_dict(ablob) bdict = blob_to_dict(bblob) pkgname = os.path.basename(path) defaultvals = {} defaultvals['PKG'] = pkgname defaultvals['PKGE'] = '0' changes = [] keys = list(set(adict.keys()) | set(bdict.keys()) | set(defaultval_map.keys())) for key in keys: astr = adict.get(key, '') bstr = bdict.get(key, '') if key in ver_monitor_fields: monitored = report_ver or astr or bstr else: monitored = key in monitor_fields mapped_key = defaultval_map.get(key, '') if mapped_key: if not astr: astr = '%s [default]' % adict.get(mapped_key, defaultvals.get(key, '')) if not bstr: bstr = '%s [default]' % bdict.get(mapped_key, defaultvals.get(key, '')) if astr != bstr: if (not report_all) and key in numeric_fields: aval = int(astr or 0) bval = int(bstr or 0) if aval != 0: percentchg = ((bval - aval) / float(aval)) * 100 else: percentchg = 100 if abs(percentchg) < monitor_numeric_threshold: continue elif (not report_all) and key in list_fields: if key == "FILELIST" and path.endswith("-dbg") and bstr.strip() != '': continue if key in ['RPROVIDES', 'RDEPENDS', 'RRECOMMENDS', 'RSUGGESTS', 'RREPLACES', 'RCONFLICTS']: (depvera, depverb) = compare_pkg_lists(astr, bstr) if depvera == depverb: continue alist = astr.split() alist.sort() blist = bstr.split() blist.sort() # We don't care about the removal of self-dependencies if pkgname in alist and not pkgname in blist: alist.remove(pkgname) if ' '.join(alist) == ' '.join(blist): continue chg = ChangeRecord(path, key, astr, bstr, monitored) changes.append(chg) return changes def process_changes(repopath, revision1, revision2='HEAD', report_all=False, report_ver=False): repo = git.Repo(repopath) assert repo.bare == False commit = repo.commit(revision1) diff = commit.diff(revision2) changes = [] for d in diff.iter_change_type('M'): path = os.path.dirname(d.a_blob.path) if path.startswith('packages/'): filename = os.path.basename(d.a_blob.path) if filename == 'latest': changes.extend(compare_dict_blobs(path, d.a_blob, d.b_blob, report_all, report_ver)) elif filename.startswith('latest.'): chg = ChangeRecord(path, filename, d.a_blob.data_stream.read(), d.b_blob.data_stream.read(), True) changes.append(chg) elif path.startswith('images/'): filename = os.path.basename(d.a_blob.path) if filename in img_monitor_files: if filename == 'files-in-image.txt': alines = d.a_blob.data_stream.read().splitlines() blines = d.b_blob.data_stream.read().splitlines() filechanges = compare_file_lists(alines,blines) if filechanges: chg = ChangeRecord(path, filename, None, None, True) chg.filechanges = filechanges changes.append(chg) elif filename == 'installed-package-names.txt': alines = d.a_blob.data_stream.read().splitlines() blines = d.b_blob.data_stream.read().splitlines() filechanges = compare_lists(alines,blines) if filechanges: chg = ChangeRecord(path, filename, None, None, True) chg.filechanges = filechanges changes.append(chg) else: chg = ChangeRecord(path, filename, d.a_blob.data_stream.read(), d.b_blob.data_stream.read(), True) changes.append(chg) elif filename == 'image-info.txt': changes.extend(compare_dict_blobs(path, d.a_blob, d.b_blob, report_all, report_ver)) elif '/image-files/' in path: chg = ChangeRecord(path, filename, d.a_blob.data_stream.read(), d.b_blob.data_stream.read(), True) changes.append(chg) # Look for added preinst/postinst/prerm/postrm # (without reporting newly added recipes) addedpkgs = [] addedchanges = [] for d in diff.iter_change_type('A'): path = os.path.dirname(d.b_blob.path) if path.startswith('packages/'): filename = os.path.basename(d.b_blob.path) if filename == 'latest': addedpkgs.append(path) elif filename.startswith('latest.'): chg = ChangeRecord(path, filename[7:], '', d.b_blob.data_stream.read(), True) addedchanges.append(chg) for chg in addedchanges: found = False for pkg in addedpkgs: if chg.path.startswith(pkg): found = True break if not found: changes.append(chg) # Look for cleared preinst/postinst/prerm/postrm for d in diff.iter_change_type('D'): path = os.path.dirname(d.a_blob.path) if path.startswith('packages/'): filename = os.path.basename(d.a_blob.path) if filename != 'latest' and filename.startswith('latest.'): chg = ChangeRecord(path, filename[7:], d.a_blob.data_stream.read(), '', True) changes.append(chg) # Link related changes for chg in changes: if chg.monitored: for chg2 in changes: # (Check dirname in the case of fields from recipe info files) if chg.path == chg2.path or os.path.dirname(chg.path) == chg2.path: if chg2.fieldname in related_fields.get(chg.fieldname, []): chg.related.append(chg2) elif chg.path == chg2.path and chg.path.startswith('packages/') and chg2.fieldname in ['PE', 'PV', 'PR']: chg.related.append(chg2) if report_all: return changes else: return [chg for chg in changes if chg.monitored] ########NEW FILE######## __FILENAME__ = cachedpath # # Based on standard python library functions but avoid # repeated stat calls. Its assumed the files will not change from under us # so we can cache stat calls. # import os import errno import stat as statmod class CachedPath(object): def __init__(self): self.statcache = {} self.lstatcache = {} self.normpathcache = {} return def updatecache(self, x): x = self.normpath(x) if x in self.statcache: del self.statcache[x] if x in self.lstatcache: del self.lstatcache[x] def normpath(self, path): if path in self.normpathcache: return self.normpathcache[path] newpath = os.path.normpath(path) self.normpathcache[path] = newpath return newpath def _callstat(self, path): if path in self.statcache: return self.statcache[path] try: st = os.stat(path) self.statcache[path] = st return st except os.error: self.statcache[path] = False return False # We might as well call lstat and then only # call stat as well in the symbolic link case # since this turns out to be much more optimal # in real world usage of this cache def callstat(self, path): path = self.normpath(path) self.calllstat(path) return self.statcache[path] def calllstat(self, path): path = self.normpath(path) if path in self.lstatcache: return self.lstatcache[path] #bb.error("LStatpath:" + path) try: lst = os.lstat(path) self.lstatcache[path] = lst if not statmod.S_ISLNK(lst.st_mode): self.statcache[path] = lst else: self._callstat(path) return lst except (os.error, AttributeError): self.lstatcache[path] = False self.statcache[path] = False return False # This follows symbolic links, so both islink() and isdir() can be true # for the same path ono systems that support symlinks def isfile(self, path): """Test whether a path is a regular file""" st = self.callstat(path) if not st: return False return statmod.S_ISREG(st.st_mode) # Is a path a directory? # This follows symbolic links, so both islink() and isdir() # can be true for the same path on systems that support symlinks def isdir(self, s): """Return true if the pathname refers to an existing directory.""" st = self.callstat(s) if not st: return False return statmod.S_ISDIR(st.st_mode) def islink(self, path): """Test whether a path is a symbolic link""" st = self.calllstat(path) if not st: return False return statmod.S_ISLNK(st.st_mode) # Does a path exist? # This is false for dangling symbolic links on systems that support them. def exists(self, path): """Test whether a path exists. Returns False for broken symbolic links""" if self.callstat(path): return True return False def lexists(self, path): """Test whether a path exists. Returns True for broken symbolic links""" if self.calllstat(path): return True return False def stat(self, path): return self.callstat(path) def lstat(self, path): return self.calllstat(path) def walk(self, top, topdown=True, onerror=None, followlinks=False): # Matches os.walk, not os.path.walk() # We may not have read permission for top, in which case we can't # get a list of the files the directory contains. os.path.walk # always suppressed the exception then, rather than blow up for a # minor reason when (say) a thousand readable directories are still # left to visit. That logic is copied here. try: names = os.listdir(top) except os.error as err: if onerror is not None: onerror(err) return dirs, nondirs = [], [] for name in names: if self.isdir(os.path.join(top, name)): dirs.append(name) else: nondirs.append(name) if topdown: yield top, dirs, nondirs for name in dirs: new_path = os.path.join(top, name) if followlinks or not self.islink(new_path): for x in self.walk(new_path, topdown, onerror, followlinks): yield x if not topdown: yield top, dirs, nondirs ## realpath() related functions def __is_path_below(self, file, root): return (file + os.path.sep).startswith(root) def __realpath_rel(self, start, rel_path, root, loop_cnt, assume_dir): """Calculates real path of symlink 'start' + 'rel_path' below 'root'; no part of 'start' below 'root' must contain symlinks. """ have_dir = True for d in rel_path.split(os.path.sep): if not have_dir and not assume_dir: raise OSError(errno.ENOENT, "no such directory %s" % start) if d == os.path.pardir: # '..' if len(start) >= len(root): # do not follow '..' before root start = os.path.dirname(start) else: # emit warning? pass else: (start, have_dir) = self.__realpath(os.path.join(start, d), root, loop_cnt, assume_dir) assert(self.__is_path_below(start, root)) return start def __realpath(self, file, root, loop_cnt, assume_dir): while self.islink(file) and len(file) >= len(root): if loop_cnt == 0: raise OSError(errno.ELOOP, file) loop_cnt -= 1 target = os.path.normpath(os.readlink(file)) if not os.path.isabs(target): tdir = os.path.dirname(file) assert(self.__is_path_below(tdir, root)) else: tdir = root file = self.__realpath_rel(tdir, target, root, loop_cnt, assume_dir) try: is_dir = self.isdir(file) except: is_dir = False return (file, is_dir) def realpath(self, file, root, use_physdir = True, loop_cnt = 100, assume_dir = False): """ Returns the canonical path of 'file' with assuming a toplevel 'root' directory. When 'use_physdir' is set, all preceding path components of 'file' will be resolved first; this flag should be set unless it is guaranteed that there is no symlink in the path. When 'assume_dir' is not set, missing path components will raise an ENOENT error""" root = os.path.normpath(root) file = os.path.normpath(file) if not root.endswith(os.path.sep): # letting root end with '/' makes some things easier root = root + os.path.sep if not self.__is_path_below(file, root): raise OSError(errno.EINVAL, "file '%s' is not below root" % file) try: if use_physdir: file = self.__realpath_rel(root, file[(len(root) - 1):], root, loop_cnt, assume_dir) else: file = self.__realpath(file, root, loop_cnt, assume_dir)[0] except OSError as e: if e.errno == errno.ELOOP: # make ELOOP more readable; without catching it, there will # be printed a backtrace with 100s of OSError exceptions # else raise OSError(errno.ELOOP, "too much recursions while resolving '%s'; loop in '%s'" % (file, e.strerror)) raise return file ########NEW FILE######## __FILENAME__ = classextend class ClassExtender(object): def __init__(self, extname, d): self.extname = extname self.d = d self.pkgs_mapping = [] def extend_name(self, name): if name.startswith("kernel-") or name == "virtual/kernel": return name if name.startswith("rtld"): return name if name.endswith("-" + self.extname): name = name.replace("-" + self.extname, "") if name.startswith("virtual/"): subs = name.split("/", 1)[1] if not subs.startswith(self.extname): return "virtual/" + self.extname + "-" + subs return name if not name.startswith(self.extname): return self.extname + "-" + name return name def map_variable(self, varname, setvar = True): var = self.d.getVar(varname, True) if not var: return "" var = var.split() newvar = [] for v in var: newvar.append(self.extend_name(v)) newdata = " ".join(newvar) if setvar: self.d.setVar(varname, newdata) return newdata def map_regexp_variable(self, varname, setvar = True): var = self.d.getVar(varname, True) if not var: return "" var = var.split() newvar = [] for v in var: if v.startswith("^" + self.extname): newvar.append(v) elif v.startswith("^"): newvar.append("^" + self.extname + "-" + v[1:]) else: newvar.append(self.extend_name(v)) newdata = " ".join(newvar) if setvar: self.d.setVar(varname, newdata) return newdata def map_depends(self, dep): if dep.endswith(("-native", "-native-runtime")) or ('nativesdk-' in dep) or ('cross-canadian' in dep) or ('-crosssdk-' in dep): return dep else: return self.extend_name(dep) def map_depends_variable(self, varname, suffix = ""): if suffix: varname = varname + "_" + suffix deps = self.d.getVar(varname, True) if not deps: return deps = bb.utils.explode_dep_versions2(deps) newdeps = {} for dep in deps: newdeps[self.map_depends(dep)] = deps[dep] self.d.setVar(varname, bb.utils.join_deps(newdeps, False)) def map_packagevars(self): for pkg in (self.d.getVar("PACKAGES", True).split() + [""]): self.map_depends_variable("RDEPENDS", pkg) self.map_depends_variable("RRECOMMENDS", pkg) self.map_depends_variable("RSUGGESTS", pkg) self.map_depends_variable("RPROVIDES", pkg) self.map_depends_variable("RREPLACES", pkg) self.map_depends_variable("RCONFLICTS", pkg) self.map_depends_variable("PKG", pkg) def rename_packages(self): for pkg in (self.d.getVar("PACKAGES", True) or "").split(): if pkg.startswith(self.extname): self.pkgs_mapping.append([pkg.split(self.extname + "-")[1], pkg]) continue self.pkgs_mapping.append([pkg, self.extend_name(pkg)]) self.d.setVar("PACKAGES", " ".join([row[1] for row in self.pkgs_mapping])) def rename_package_variables(self, variables): for pkg_mapping in self.pkgs_mapping: for subs in variables: self.d.renameVar("%s_%s" % (subs, pkg_mapping[0]), "%s_%s" % (subs, pkg_mapping[1])) class NativesdkClassExtender(ClassExtender): def map_depends(self, dep): if dep.endswith(("-gcc-initial", "-gcc", "-g++")): return dep + "-crosssdk" elif dep.endswith(("-native", "-native-runtime")) or ('nativesdk-' in dep) or ('-cross-' in dep) or ('-crosssdk-' in dep): return dep else: return self.extend_name(dep) ########NEW FILE######## __FILENAME__ = classutils class ClassRegistry(type): """Maintain a registry of classes, indexed by name. Note that this implementation requires that the names be unique, as it uses a dictionary to hold the classes by name. The name in the registry can be overridden via the 'name' attribute of the class, and the 'priority' attribute controls priority. The prioritized() method returns the registered classes in priority order. Subclasses of ClassRegistry may define an 'implemented' property to exert control over whether the class will be added to the registry (e.g. to keep abstract base classes out of the registry).""" priority = 0 class __metaclass__(type): """Give each ClassRegistry their own registry""" def __init__(cls, name, bases, attrs): cls.registry = {} type.__init__(cls, name, bases, attrs) def __init__(cls, name, bases, attrs): super(ClassRegistry, cls).__init__(name, bases, attrs) try: if not cls.implemented: return except AttributeError: pass try: cls.name except AttributeError: cls.name = name cls.registry[cls.name] = cls @classmethod def prioritized(tcls): return sorted(tcls.registry.values(), key=lambda v: v.priority, reverse=True) def unregister(cls): for key in cls.registry.keys(): if cls.registry[key] is cls: del cls.registry[key] ########NEW FILE######## __FILENAME__ = data import oe.maketype def typed_value(key, d): """Construct a value for the specified metadata variable, using its flags to determine the type and parameters for construction.""" var_type = d.getVarFlag(key, 'type') flags = d.getVarFlags(key) if flags is not None: flags = dict((flag, d.expand(value)) for flag, value in flags.iteritems()) else: flags = {} try: return oe.maketype.create(d.getVar(key, True) or '', var_type, **flags) except (TypeError, ValueError), exc: bb.msg.fatal("Data", "%s: %s" % (key, str(exc))) ########NEW FILE######## __FILENAME__ = distro_check def get_links_from_url(url): "Return all the href links found on the web location" import urllib, sgmllib class LinksParser(sgmllib.SGMLParser): def parse(self, s): "Parse the given string 's'." self.feed(s) self.close() def __init__(self, verbose=0): "Initialise an object passing 'verbose' to the superclass." sgmllib.SGMLParser.__init__(self, verbose) self.hyperlinks = [] def start_a(self, attributes): "Process a hyperlink and its 'attributes'." for name, value in attributes: if name == "href": self.hyperlinks.append(value.strip('/')) def get_hyperlinks(self): "Return the list of hyperlinks." return self.hyperlinks sock = urllib.urlopen(url) webpage = sock.read() sock.close() linksparser = LinksParser() linksparser.parse(webpage) return linksparser.get_hyperlinks() def find_latest_numeric_release(url): "Find the latest listed numeric release on the given url" max=0 maxstr="" for link in get_links_from_url(url): try: release = float(link) except: release = 0 if release > max: max = release maxstr = link return maxstr def is_src_rpm(name): "Check if the link is pointing to a src.rpm file" if name[-8:] == ".src.rpm": return True else: return False def package_name_from_srpm(srpm): "Strip out the package name from the src.rpm filename" strings = srpm.split('-') package_name = strings[0] for i in range(1, len (strings) - 1): str = strings[i] if not str[0].isdigit(): package_name += '-' + str return package_name def clean_package_list(package_list): "Removes multiple entries of packages and sorts the list" set = {} map(set.__setitem__, package_list, []) return set.keys() def get_latest_released_meego_source_package_list(): "Returns list of all the name os packages in the latest meego distro" package_names = [] try: f = open("/tmp/Meego-1.1", "r") for line in f: package_names.append(line[:-1] + ":" + "main") # Also strip the '\n' at the end except IOError: pass package_list=clean_package_list(package_names) return "1.0", package_list def get_source_package_list_from_url(url, section): "Return a sectioned list of package names from a URL list" bb.note("Reading %s: %s" % (url, section)) links = get_links_from_url(url) srpms = filter(is_src_rpm, links) names_list = map(package_name_from_srpm, srpms) new_pkgs = [] for pkgs in names_list: new_pkgs.append(pkgs + ":" + section) return new_pkgs def get_latest_released_fedora_source_package_list(): "Returns list of all the name os packages in the latest fedora distro" latest = find_latest_numeric_release("http://archive.fedoraproject.org/pub/fedora/linux/releases/") package_names = get_source_package_list_from_url("http://archive.fedoraproject.org/pub/fedora/linux/releases/%s/Fedora/source/SRPMS/" % latest, "main") # package_names += get_source_package_list_from_url("http://download.fedora.redhat.com/pub/fedora/linux/releases/%s/Everything/source/SPRMS/" % latest, "everything") package_names += get_source_package_list_from_url("http://archive.fedoraproject.org/pub/fedora/linux/updates/%s/SRPMS/" % latest, "updates") package_list=clean_package_list(package_names) return latest, package_list def get_latest_released_opensuse_source_package_list(): "Returns list of all the name os packages in the latest opensuse distro" latest = find_latest_numeric_release("http://download.opensuse.org/source/distribution/") package_names = get_source_package_list_from_url("http://download.opensuse.org/source/distribution/%s/repo/oss/suse/src/" % latest, "main") package_names += get_source_package_list_from_url("http://download.opensuse.org/update/%s/rpm/src/" % latest, "updates") package_list=clean_package_list(package_names) return latest, package_list def get_latest_released_mandriva_source_package_list(): "Returns list of all the name os packages in the latest mandriva distro" latest = find_latest_numeric_release("http://distrib-coffee.ipsl.jussieu.fr/pub/linux/MandrivaLinux/official/") package_names = get_source_package_list_from_url("http://distrib-coffee.ipsl.jussieu.fr/pub/linux/MandrivaLinux/official/%s/SRPMS/main/release/" % latest, "main") # package_names += get_source_package_list_from_url("http://distrib-coffee.ipsl.jussieu.fr/pub/linux/MandrivaLinux/official/%s/SRPMS/contrib/release/" % latest, "contrib") package_names += get_source_package_list_from_url("http://distrib-coffee.ipsl.jussieu.fr/pub/linux/MandrivaLinux/official/%s/SRPMS/main/updates/" % latest, "updates") package_list=clean_package_list(package_names) return latest, package_list def find_latest_debian_release(url): "Find the latest listed debian release on the given url" releases = [] for link in get_links_from_url(url): if link[:6] == "Debian": if ';' not in link: releases.append(link) releases.sort() try: return releases.pop()[6:] except: return "_NotFound_" def get_debian_style_source_package_list(url, section): "Return the list of package-names stored in the debian style Sources.gz file" import urllib sock = urllib.urlopen(url) import tempfile tmpfile = tempfile.NamedTemporaryFile(mode='wb', prefix='oecore.', suffix='.tmp', delete=False) tmpfilename=tmpfile.name tmpfile.write(sock.read()) sock.close() tmpfile.close() import gzip bb.note("Reading %s: %s" % (url, section)) f = gzip.open(tmpfilename) package_names = [] for line in f: if line[:9] == "Package: ": package_names.append(line[9:-1] + ":" + section) # Also strip the '\n' at the end os.unlink(tmpfilename) return package_names def get_latest_released_debian_source_package_list(): "Returns list of all the name os packages in the latest debian distro" latest = find_latest_debian_release("http://ftp.debian.org/debian/dists/") url = "http://ftp.debian.org/debian/dists/stable/main/source/Sources.gz" package_names = get_debian_style_source_package_list(url, "main") # url = "http://ftp.debian.org/debian/dists/stable/contrib/source/Sources.gz" # package_names += get_debian_style_source_package_list(url, "contrib") url = "http://ftp.debian.org/debian/dists/stable-proposed-updates/main/source/Sources.gz" package_names += get_debian_style_source_package_list(url, "updates") package_list=clean_package_list(package_names) return latest, package_list def find_latest_ubuntu_release(url): "Find the latest listed ubuntu release on the given url" url += "?C=M;O=D" # Descending Sort by Last Modified for link in get_links_from_url(url): if link[-8:] == "-updates": return link[:-8] return "_NotFound_" def get_latest_released_ubuntu_source_package_list(): "Returns list of all the name os packages in the latest ubuntu distro" latest = find_latest_ubuntu_release("http://archive.ubuntu.com/ubuntu/dists/") url = "http://archive.ubuntu.com/ubuntu/dists/%s/main/source/Sources.gz" % latest package_names = get_debian_style_source_package_list(url, "main") # url = "http://archive.ubuntu.com/ubuntu/dists/%s/multiverse/source/Sources.gz" % latest # package_names += get_debian_style_source_package_list(url, "multiverse") # url = "http://archive.ubuntu.com/ubuntu/dists/%s/universe/source/Sources.gz" % latest # package_names += get_debian_style_source_package_list(url, "universe") url = "http://archive.ubuntu.com/ubuntu/dists/%s-updates/main/source/Sources.gz" % latest package_names += get_debian_style_source_package_list(url, "updates") package_list=clean_package_list(package_names) return latest, package_list def create_distro_packages_list(distro_check_dir): pkglst_dir = os.path.join(distro_check_dir, "package_lists") if not os.path.isdir (pkglst_dir): os.makedirs(pkglst_dir) # first clear old stuff for file in os.listdir(pkglst_dir): os.unlink(os.path.join(pkglst_dir, file)) per_distro_functions = [ ["Debian", get_latest_released_debian_source_package_list], ["Ubuntu", get_latest_released_ubuntu_source_package_list], ["Fedora", get_latest_released_fedora_source_package_list], ["OpenSuSE", get_latest_released_opensuse_source_package_list], ["Mandriva", get_latest_released_mandriva_source_package_list], ["Meego", get_latest_released_meego_source_package_list] ] from datetime import datetime begin = datetime.now() for distro in per_distro_functions: name = distro[0] release, package_list = distro[1]() bb.note("Distro: %s, Latest Release: %s, # src packages: %d" % (name, release, len(package_list))) package_list_file = os.path.join(pkglst_dir, name + "-" + release) f = open(package_list_file, "w+b") for pkg in package_list: f.write(pkg + "\n") f.close() end = datetime.now() delta = end - begin bb.note("package_list generatiosn took this much time: %d seconds" % delta.seconds) def update_distro_data(distro_check_dir, datetime): """ If distro packages list data is old then rebuild it. The operations has to be protected by a lock so that only one thread performes it at a time. """ if not os.path.isdir (distro_check_dir): try: bb.note ("Making new directory: %s" % distro_check_dir) os.makedirs (distro_check_dir) except OSError: raise Exception('Unable to create directory %s' % (distro_check_dir)) datetime_file = os.path.join(distro_check_dir, "build_datetime") saved_datetime = "_invalid_" import fcntl try: if not os.path.exists(datetime_file): open(datetime_file, 'w+b').close() # touch the file so that the next open won't fail f = open(datetime_file, "r+b") fcntl.lockf(f, fcntl.LOCK_EX) saved_datetime = f.read() if saved_datetime[0:8] != datetime[0:8]: bb.note("The build datetime did not match: saved:%s current:%s" % (saved_datetime, datetime)) bb.note("Regenerating distro package lists") create_distro_packages_list(distro_check_dir) f.seek(0) f.write(datetime) except OSError: raise Exception('Unable to read/write this file: %s' % (datetime_file)) finally: fcntl.lockf(f, fcntl.LOCK_UN) f.close() def compare_in_distro_packages_list(distro_check_dir, d): if not os.path.isdir(distro_check_dir): raise Exception("compare_in_distro_packages_list: invalid distro_check_dir passed") localdata = bb.data.createCopy(d) pkglst_dir = os.path.join(distro_check_dir, "package_lists") matching_distros = [] pn = d.getVar('PN', True) recipe_name = d.getVar('PN', True) bb.note("Checking: %s" % pn) trim_dict = dict({"-native":"-native", "-cross":"-cross", "-initial":"-initial"}) if pn.find("-native") != -1: pnstripped = pn.split("-native") localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True)) bb.data.update_data(localdata) recipe_name = pnstripped[0] if pn.startswith("nativesdk-"): pnstripped = pn.split("nativesdk-") localdata.setVar('OVERRIDES', "pn-" + pnstripped[1] + ":" + d.getVar('OVERRIDES', True)) bb.data.update_data(localdata) recipe_name = pnstripped[1] if pn.find("-cross") != -1: pnstripped = pn.split("-cross") localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True)) bb.data.update_data(localdata) recipe_name = pnstripped[0] if pn.find("-initial") != -1: pnstripped = pn.split("-initial") localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True)) bb.data.update_data(localdata) recipe_name = pnstripped[0] bb.note("Recipe: %s" % recipe_name) tmp = localdata.getVar('DISTRO_PN_ALIAS', True) distro_exceptions = dict({"OE-Core":'OE-Core', "OpenedHand":'OpenedHand', "Intel":'Intel', "Upstream":'Upstream', "Windriver":'Windriver', "OSPDT":'OSPDT Approved', "Poky":'poky'}) if tmp: list = tmp.split(' ') for str in list: if str and str.find("=") == -1 and distro_exceptions[str]: matching_distros.append(str) distro_pn_aliases = {} if tmp: list = tmp.split(' ') for str in list: if str.find("=") != -1: (dist, pn_alias) = str.split('=') distro_pn_aliases[dist.strip().lower()] = pn_alias.strip() for file in os.listdir(pkglst_dir): (distro, distro_release) = file.split("-") f = open(os.path.join(pkglst_dir, file), "rb") for line in f: (pkg, section) = line.split(":") if distro.lower() in distro_pn_aliases: pn = distro_pn_aliases[distro.lower()] else: pn = recipe_name if pn == pkg: matching_distros.append(distro + "-" + section[:-1]) # strip the \n at the end f.close() break f.close() if tmp != None: list = tmp.split(' ') for item in list: matching_distros.append(item) bb.note("Matching: %s" % matching_distros) return matching_distros def create_log_file(d, logname): import subprocess logpath = d.getVar('LOG_DIR', True) bb.utils.mkdirhier(logpath) logfn, logsuffix = os.path.splitext(logname) logfile = os.path.join(logpath, "%s.%s%s" % (logfn, d.getVar('DATETIME', True), logsuffix)) if not os.path.exists(logfile): slogfile = os.path.join(logpath, logname) if os.path.exists(slogfile): os.remove(slogfile) subprocess.call("touch %s" % logfile, shell=True) os.symlink(logfile, slogfile) d.setVar('LOG_FILE', logfile) return logfile def save_distro_check_result(result, datetime, result_file, d): pn = d.getVar('PN', True) logdir = d.getVar('LOG_DIR', True) if not logdir: bb.error("LOG_DIR variable is not defined, can't write the distro_check results") return if not os.path.isdir(logdir): os.makedirs(logdir) line = pn for i in result: line = line + "," + i f = open(result_file, "a") import fcntl fcntl.lockf(f, fcntl.LOCK_EX) f.seek(0, os.SEEK_END) # seek to the end of file f.write(line + "\n") fcntl.lockf(f, fcntl.LOCK_UN) f.close() ########NEW FILE######## __FILENAME__ = image from oe.utils import execute_pre_post_process import os import subprocess import multiprocessing def generate_image(arg): (type, subimages, create_img_cmd) = arg bb.note("Running image creation script for %s: %s ..." % (type, create_img_cmd)) try: subprocess.check_output(create_img_cmd, stderr=subprocess.STDOUT) except subprocess.CalledProcessError as e: return("Error: The image creation script '%s' returned %d:\n%s" % (e.cmd, e.returncode, e.output)) return None """ This class will help compute IMAGE_FSTYPE dependencies and group them in batches that can be executed in parallel. The next example is for illustration purposes, highly unlikely to happen in real life. It's just one of the test cases I used to test the algorithm: For: IMAGE_FSTYPES = "i1 i2 i3 i4 i5" IMAGE_TYPEDEP_i4 = "i2" IMAGE_TYPEDEP_i5 = "i6 i4" IMAGE_TYPEDEP_i6 = "i7" IMAGE_TYPEDEP_i7 = "i2" We get the following list of batches that can be executed in parallel, having the dependencies satisfied: [['i1', 'i3', 'i2'], ['i4', 'i7'], ['i6'], ['i5']] """ class ImageDepGraph(object): def __init__(self, d): self.d = d self.graph = dict() self.deps_array = dict() def _construct_dep_graph(self, image_fstypes): graph = dict() def add_node(node): deps = (self.d.getVar('IMAGE_TYPEDEP_' + node, True) or "") if deps != "": graph[node] = deps for dep in deps.split(): if not dep in graph: add_node(dep) else: graph[node] = "" for fstype in image_fstypes: add_node(fstype) return graph def _clean_graph(self): # Live and VMDK images will be processed via inheriting # bbclass and does not get processed here. Remove them from the fstypes # graph. Their dependencies are already added, so no worries here. remove_list = (self.d.getVar('IMAGE_TYPES_MASKED', True) or "").split() for item in remove_list: self.graph.pop(item, None) def _compute_dependencies(self): """ returns dict object of nodes with [no_of_depends_on, no_of_depended_by] for each node """ deps_array = dict() for node in self.graph: deps_array[node] = [0, 0] for node in self.graph: deps = self.graph[node].split() deps_array[node][0] += len(deps) for dep in deps: deps_array[dep][1] += 1 return deps_array def _sort_graph(self): sorted_list = [] group = [] for node in self.graph: if node not in self.deps_array: continue depends_on = self.deps_array[node][0] if depends_on == 0: group.append(node) if len(group) == 0 and len(self.deps_array) != 0: bb.fatal("possible fstype circular dependency...") sorted_list.append(group) # remove added nodes from deps_array for item in group: for node in self.graph: if item in self.graph[node]: self.deps_array[node][0] -= 1 self.deps_array.pop(item, None) if len(self.deps_array): # recursive call, to find the next group sorted_list += self._sort_graph() return sorted_list def group_fstypes(self, image_fstypes): self.graph = self._construct_dep_graph(image_fstypes) self._clean_graph() self.deps_array = self._compute_dependencies() alltypes = [node for node in self.graph] return (alltypes, self._sort_graph()) class Image(ImageDepGraph): def __init__(self, d): self.d = d super(Image, self).__init__(d) def _get_rootfs_size(self): """compute the rootfs size""" rootfs_alignment = int(self.d.getVar('IMAGE_ROOTFS_ALIGNMENT', True)) overhead_factor = float(self.d.getVar('IMAGE_OVERHEAD_FACTOR', True)) rootfs_req_size = int(self.d.getVar('IMAGE_ROOTFS_SIZE', True)) rootfs_extra_space = eval(self.d.getVar('IMAGE_ROOTFS_EXTRA_SPACE', True)) output = subprocess.check_output(['du', '-ks', self.d.getVar('IMAGE_ROOTFS', True)]) size_kb = int(output.split()[0]) base_size = size_kb * overhead_factor base_size = (base_size, rootfs_req_size)[base_size < rootfs_req_size] + \ rootfs_extra_space if base_size != int(base_size): base_size = int(base_size + 1) base_size += rootfs_alignment - 1 base_size -= base_size % rootfs_alignment return base_size def _create_symlinks(self, subimages): """create symlinks to the newly created image""" deploy_dir = self.d.getVar('DEPLOY_DIR_IMAGE', True) img_name = self.d.getVar('IMAGE_NAME', True) link_name = self.d.getVar('IMAGE_LINK_NAME', True) manifest_name = self.d.getVar('IMAGE_MANIFEST', True) os.chdir(deploy_dir) if link_name is not None: for type in subimages: if os.path.exists(img_name + ".rootfs." + type): dst = link_name + "." + type src = img_name + ".rootfs." + type bb.note("Creating symlink: %s -> %s" % (dst, src)) os.symlink(src, dst) if manifest_name is not None and \ os.path.exists(manifest_name) and \ not os.path.exists(link_name + ".manifest"): os.symlink(os.path.basename(manifest_name), link_name + ".manifest") def _remove_old_symlinks(self): """remove the symlinks to old binaries""" if self.d.getVar('IMAGE_LINK_NAME', True): deploy_dir = self.d.getVar('DEPLOY_DIR_IMAGE', True) for img in os.listdir(deploy_dir): if img.find(self.d.getVar('IMAGE_LINK_NAME', True)) == 0: img = os.path.join(deploy_dir, img) if os.path.islink(img): if self.d.getVar('RM_OLD_IMAGE', True) == "1" and \ os.path.exists(os.path.realpath(img)): os.remove(os.path.realpath(img)) os.remove(img) """ This function will just filter out the compressed image types from the fstype groups returning a (filtered_fstype_groups, cimages) tuple. """ def _filter_out_commpressed(self, fstype_groups): ctypes = self.d.getVar('COMPRESSIONTYPES', True).split() cimages = {} filtered_groups = [] for group in fstype_groups: filtered_group = [] for type in group: basetype = None for ctype in ctypes: if type.endswith("." + ctype): basetype = type[:-len("." + ctype)] if basetype not in filtered_group: filtered_group.append(basetype) if basetype not in cimages: cimages[basetype] = [] if ctype not in cimages[basetype]: cimages[basetype].append(ctype) break if not basetype and type not in filtered_group: filtered_group.append(type) filtered_groups.append(filtered_group) return (filtered_groups, cimages) def _get_image_types(self): """returns a (types, cimages) tuple""" alltypes, fstype_groups = self.group_fstypes(self.d.getVar('IMAGE_FSTYPES', True).split()) filtered_groups, cimages = self._filter_out_commpressed(fstype_groups) return (alltypes, filtered_groups, cimages) def _write_script(self, type, cmds): tempdir = self.d.getVar('T', True) script_name = os.path.join(tempdir, "create_image." + type) self.d.setVar('img_creation_func', '\n'.join(cmds)) self.d.setVarFlag('img_creation_func', 'func', 1) self.d.setVarFlag('img_creation_func', 'fakeroot', 1) with open(script_name, "w+") as script: script.write("%s" % bb.build.shell_trap_code()) script.write("export ROOTFS_SIZE=%d\n" % self._get_rootfs_size()) bb.data.emit_func('img_creation_func', script, self.d) script.write("img_creation_func\n") os.chmod(script_name, 0775) return script_name def _get_imagecmds(self): old_overrides = self.d.getVar('OVERRIDES', 0) alltypes, fstype_groups, cimages = self._get_image_types() image_cmd_groups = [] bb.note("The image creation groups are: %s" % str(fstype_groups)) for fstype_group in fstype_groups: image_cmds = [] for type in fstype_group: cmds = [] subimages = [] localdata = bb.data.createCopy(self.d) localdata.setVar('OVERRIDES', '%s:%s' % (type, old_overrides)) bb.data.update_data(localdata) localdata.setVar('type', type) cmds.append("\t" + localdata.getVar("IMAGE_CMD", True)) cmds.append(localdata.expand("\tcd ${DEPLOY_DIR_IMAGE}")) if type in cimages: for ctype in cimages[type]: cmds.append("\t" + localdata.getVar("COMPRESS_CMD_" + ctype, True)) subimages.append(type + "." + ctype) if type not in alltypes: cmds.append(localdata.expand("\trm ${IMAGE_NAME}.rootfs.${type}")) else: subimages.append(type) script_name = self._write_script(type, cmds) image_cmds.append((type, subimages, script_name)) image_cmd_groups.append(image_cmds) return image_cmd_groups def create(self): bb.note("###### Generate images #######") pre_process_cmds = self.d.getVar("IMAGE_PREPROCESS_COMMAND", True) post_process_cmds = self.d.getVar("IMAGE_POSTPROCESS_COMMAND", True) execute_pre_post_process(self.d, pre_process_cmds) self._remove_old_symlinks() image_cmd_groups = self._get_imagecmds() for image_cmds in image_cmd_groups: # create the images in parallel nproc = multiprocessing.cpu_count() pool = bb.utils.multiprocessingpool(nproc) results = list(pool.imap(generate_image, image_cmds)) pool.close() pool.join() for result in results: if result is not None: bb.fatal(result) for image_type, subimages, script in image_cmds: bb.note("Creating symlinks for %s image ..." % image_type) self._create_symlinks(subimages) execute_pre_post_process(self.d, post_process_cmds) def create_image(d): Image(d).create() if __name__ == "__main__": """ Image creation can be called independent from bitbake environment. """ """ TBD """ ########NEW FILE######## __FILENAME__ = license # vi:sts=4:sw=4:et """Code for parsing OpenEmbedded license strings""" import ast import re from fnmatch import fnmatchcase as fnmatch class LicenseError(Exception): pass class LicenseSyntaxError(LicenseError): def __init__(self, licensestr, exc): self.licensestr = licensestr self.exc = exc LicenseError.__init__(self) def __str__(self): return "error in '%s': %s" % (self.licensestr, self.exc) class InvalidLicense(LicenseError): def __init__(self, license): self.license = license LicenseError.__init__(self) def __str__(self): return "invalid characters in license '%s'" % self.license license_operator = re.compile('([&|() ])') license_pattern = re.compile('[a-zA-Z0-9.+_\-]+$') class LicenseVisitor(ast.NodeVisitor): """Syntax tree visitor which can accept OpenEmbedded license strings""" def visit_string(self, licensestr): new_elements = [] elements = filter(lambda x: x.strip(), license_operator.split(licensestr)) for pos, element in enumerate(elements): if license_pattern.match(element): if pos > 0 and license_pattern.match(elements[pos-1]): new_elements.append('&') element = '"' + element + '"' elif not license_operator.match(element): raise InvalidLicense(element) new_elements.append(element) self.visit(ast.parse(' '.join(new_elements))) class FlattenVisitor(LicenseVisitor): """Flatten a license tree (parsed from a string) by selecting one of each set of OR options, in the way the user specifies""" def __init__(self, choose_licenses): self.choose_licenses = choose_licenses self.licenses = [] LicenseVisitor.__init__(self) def visit_Str(self, node): self.licenses.append(node.s) def visit_BinOp(self, node): if isinstance(node.op, ast.BitOr): left = FlattenVisitor(self.choose_licenses) left.visit(node.left) right = FlattenVisitor(self.choose_licenses) right.visit(node.right) selected = self.choose_licenses(left.licenses, right.licenses) self.licenses.extend(selected) else: self.generic_visit(node) def flattened_licenses(licensestr, choose_licenses): """Given a license string and choose_licenses function, return a flat list of licenses""" flatten = FlattenVisitor(choose_licenses) try: flatten.visit_string(licensestr) except SyntaxError as exc: raise LicenseSyntaxError(licensestr, exc) return flatten.licenses def is_included(licensestr, whitelist=None, blacklist=None): """Given a license string and whitelist and blacklist, determine if the license string matches the whitelist and does not match the blacklist. Returns a tuple holding the boolean state and a list of the applicable licenses which were excluded (or None, if the state is True) """ def include_license(license): return any(fnmatch(license, pattern) for pattern in whitelist) def exclude_license(license): return any(fnmatch(license, pattern) for pattern in blacklist) def choose_licenses(alpha, beta): """Select the option in an OR which is the 'best' (has the most included licenses).""" alpha_weight = len(filter(include_license, alpha)) beta_weight = len(filter(include_license, beta)) if alpha_weight > beta_weight: return alpha else: return beta if not whitelist: whitelist = ['*'] if not blacklist: blacklist = [] licenses = flattened_licenses(licensestr, choose_licenses) excluded = filter(lambda lic: exclude_license(lic), licenses) included = filter(lambda lic: include_license(lic), licenses) if excluded: return False, excluded else: return True, included ########NEW FILE######## __FILENAME__ = lsb def release_dict(): """Return the output of lsb_release -ir as a dictionary""" from subprocess import PIPE try: output, err = bb.process.run(['lsb_release', '-ir'], stderr=PIPE) except bb.process.CmdError as exc: return None data = {} for line in output.splitlines(): try: key, value = line.split(":\t", 1) except ValueError: continue else: data[key] = value return data def release_dict_file(): """ Try to gather LSB release information manually when lsb_release tool is unavailable """ data = None try: if os.path.exists('/etc/lsb-release'): data = {} with open('/etc/lsb-release') as f: for line in f: key, value = line.split("=", 1) data[key] = value.strip() elif os.path.exists('/etc/redhat-release'): data = {} with open('/etc/redhat-release') as f: distro = f.readline().strip() import re match = re.match(r'(.*) release (.*) \((.*)\)', distro) if match: data['DISTRIB_ID'] = match.group(1) data['DISTRIB_RELEASE'] = match.group(2) elif os.path.exists('/etc/SuSE-release'): data = {} data['DISTRIB_ID'] = 'SUSE LINUX' with open('/etc/SuSE-release') as f: for line in f: if line.startswith('VERSION = '): data['DISTRIB_RELEASE'] = line[10:].rstrip() break elif os.path.exists('/etc/os-release'): data = {} with open('/etc/os-release') as f: for line in f: if line.startswith('NAME='): data['DISTRIB_ID'] = line[5:].rstrip().strip('"') if line.startswith('VERSION_ID='): data['DISTRIB_RELEASE'] = line[11:].rstrip().strip('"') except IOError: return None return data def distro_identifier(adjust_hook=None): """Return a distro identifier string based upon lsb_release -ri, with optional adjustment via a hook""" lsb_data = release_dict() if lsb_data: distro_id, release = lsb_data['Distributor ID'], lsb_data['Release'] else: lsb_data_file = release_dict_file() if lsb_data_file: distro_id, release = lsb_data_file['DISTRIB_ID'], lsb_data_file.get('DISTRIB_RELEASE', None) else: distro_id, release = None, None if adjust_hook: distro_id, release = adjust_hook(distro_id, release) if not distro_id: return "Unknown" if release: id_str = '{0}-{1}'.format(distro_id, release) else: id_str = distro_id return id_str.replace(' ','-').replace('/','-') ########NEW FILE######## __FILENAME__ = maketype """OpenEmbedded variable typing support Types are defined in the metadata by name, using the 'type' flag on a variable. Other flags may be utilized in the construction of the types. See the arguments of the type's factory for details. """ import inspect import types available_types = {} class MissingFlag(TypeError): """A particular flag is required to construct the type, but has not been provided.""" def __init__(self, flag, type): self.flag = flag self.type = type TypeError.__init__(self) def __str__(self): return "Type '%s' requires flag '%s'" % (self.type, self.flag) def factory(var_type): """Return the factory for a specified type.""" if var_type is None: raise TypeError("No type specified. Valid types: %s" % ', '.join(available_types)) try: return available_types[var_type] except KeyError: raise TypeError("Invalid type '%s':\n Valid types: %s" % (var_type, ', '.join(available_types))) def create(value, var_type, **flags): """Create an object of the specified type, given the specified flags and string value.""" obj = factory(var_type) objflags = {} for flag in obj.flags: if flag not in flags: if flag not in obj.optflags: raise MissingFlag(flag, var_type) else: objflags[flag] = flags[flag] return obj(value, **objflags) def get_callable_args(obj): """Grab all but the first argument of the specified callable, returning the list, as well as a list of which of the arguments have default values.""" if type(obj) is type: obj = obj.__init__ args, varargs, keywords, defaults = inspect.getargspec(obj) flaglist = [] if args: if len(args) > 1 and args[0] == 'self': args = args[1:] flaglist.extend(args) optional = set() if defaults: optional |= set(flaglist[-len(defaults):]) return flaglist, optional def factory_setup(name, obj): """Prepare a factory for use.""" args, optional = get_callable_args(obj) extra_args = args[1:] if extra_args: obj.flags, optional = extra_args, optional obj.optflags = set(optional) else: obj.flags = obj.optflags = () if not hasattr(obj, 'name'): obj.name = name def register(name, factory): """Register a type, given its name and a factory callable. Determines the required and optional flags from the factory's arguments.""" factory_setup(name, factory) available_types[factory.name] = factory # Register all our included types for name in dir(types): if name.startswith('_'): continue obj = getattr(types, name) if not callable(obj): continue register(name, obj) ########NEW FILE######## __FILENAME__ = manifest from abc import ABCMeta, abstractmethod import os import re import bb class Manifest(object): """ This is an abstract class. Do not instantiate this directly. """ __metaclass__ = ABCMeta PKG_TYPE_MUST_INSTALL = "mip" PKG_TYPE_MULTILIB = "mlp" PKG_TYPE_LANGUAGE = "lgp" PKG_TYPE_ATTEMPT_ONLY = "aop" MANIFEST_TYPE_IMAGE = "image" MANIFEST_TYPE_SDK_HOST = "sdk_host" MANIFEST_TYPE_SDK_TARGET = "sdk_target" var_maps = { MANIFEST_TYPE_IMAGE: { "PACKAGE_INSTALL": PKG_TYPE_MUST_INSTALL, "PACKAGE_INSTALL_ATTEMPTONLY": PKG_TYPE_ATTEMPT_ONLY, "LINGUAS_INSTALL": PKG_TYPE_LANGUAGE }, MANIFEST_TYPE_SDK_HOST: { "TOOLCHAIN_HOST_TASK": PKG_TYPE_MUST_INSTALL, "TOOLCHAIN_HOST_TASK_ATTEMPTONLY": PKG_TYPE_ATTEMPT_ONLY }, MANIFEST_TYPE_SDK_TARGET: { "TOOLCHAIN_TARGET_TASK": PKG_TYPE_MUST_INSTALL, "TOOLCHAIN_TARGET_ATTEMPTONLY": PKG_TYPE_ATTEMPT_ONLY } } INSTALL_ORDER = [ PKG_TYPE_LANGUAGE, PKG_TYPE_MUST_INSTALL, PKG_TYPE_ATTEMPT_ONLY, PKG_TYPE_MULTILIB ] initial_manifest_file_header = \ "# This file was generated automatically and contains the packages\n" \ "# passed on to the package manager in order to create the rootfs.\n\n" \ "# Format:\n" \ "# <package_type>,<package_name>\n" \ "# where:\n" \ "# <package_type> can be:\n" \ "# 'mip' = must install package\n" \ "# 'aop' = attempt only package\n" \ "# 'mlp' = multilib package\n" \ "# 'lgp' = language package\n\n" def __init__(self, d, manifest_dir=None, manifest_type=MANIFEST_TYPE_IMAGE): self.d = d self.manifest_type = manifest_type if manifest_dir is None: if manifest_type != self.MANIFEST_TYPE_IMAGE: self.manifest_dir = self.d.getVar('SDK_DIR', True) else: self.manifest_dir = self.d.getVar('WORKDIR', True) else: self.manifest_dir = manifest_dir bb.utils.mkdirhier(self.manifest_dir) self.initial_manifest = os.path.join(self.manifest_dir, "%s_initial_manifest" % manifest_type) self.final_manifest = os.path.join(self.manifest_dir, "%s_final_manifest" % manifest_type) self.full_manifest = os.path.join(self.manifest_dir, "%s_full_manifest" % manifest_type) # packages in the following vars will be split in 'must install' and # 'multilib' self.vars_to_split = ["PACKAGE_INSTALL", "TOOLCHAIN_HOST_TASK", "TOOLCHAIN_TARGET_TASK"] """ This creates a standard initial manifest for core-image-(minimal|sato|sato-sdk). This will be used for testing until the class is implemented properly! """ def _create_dummy_initial(self): image_rootfs = self.d.getVar('IMAGE_ROOTFS', True) pkg_list = dict() if image_rootfs.find("core-image-sato-sdk") > 0: pkg_list[self.PKG_TYPE_MUST_INSTALL] = \ "packagegroup-core-x11-sato-games packagegroup-base-extended " \ "packagegroup-core-x11-sato packagegroup-core-x11-base " \ "packagegroup-core-sdk packagegroup-core-tools-debug " \ "packagegroup-core-boot packagegroup-core-tools-testapps " \ "packagegroup-core-eclipse-debug packagegroup-core-qt-demoapps " \ "apt packagegroup-core-tools-profile psplash " \ "packagegroup-core-standalone-sdk-target " \ "packagegroup-core-ssh-openssh dpkg kernel-dev" pkg_list[self.PKG_TYPE_LANGUAGE] = \ "locale-base-en-us locale-base-en-gb" elif image_rootfs.find("core-image-sato") > 0: pkg_list[self.PKG_TYPE_MUST_INSTALL] = \ "packagegroup-core-ssh-dropbear packagegroup-core-x11-sato-games " \ "packagegroup-core-x11-base psplash apt dpkg packagegroup-base-extended " \ "packagegroup-core-x11-sato packagegroup-core-boot" pkg_list['lgp'] = \ "locale-base-en-us locale-base-en-gb" elif image_rootfs.find("core-image-minimal") > 0: pkg_list[self.PKG_TYPE_MUST_INSTALL] = "run-postinsts packagegroup-core-boot" with open(self.initial_manifest, "w+") as manifest: manifest.write(self.initial_manifest_file_header) for pkg_type in pkg_list: for pkg in pkg_list[pkg_type].split(): manifest.write("%s,%s\n" % (pkg_type, pkg)) """ This will create the initial manifest which will be used by Rootfs class to generate the rootfs """ @abstractmethod def create_initial(self): pass """ This creates the manifest after everything has been installed. """ @abstractmethod def create_final(self): pass """ This creates the manifest after the package in initial manifest has been dummy installed. It lists all *to be installed* packages. There is no real installation, just a test. """ @abstractmethod def create_full(self, pm): pass """ The following function parses an initial manifest and returns a dictionary object with the must install, attempt only, multilib and language packages. """ def parse_initial_manifest(self): pkgs = dict() with open(self.initial_manifest) as manifest: for line in manifest.read().split('\n'): comment = re.match("^#.*", line) pattern = "^(%s|%s|%s|%s),(.*)$" % \ (self.PKG_TYPE_MUST_INSTALL, self.PKG_TYPE_ATTEMPT_ONLY, self.PKG_TYPE_MULTILIB, self.PKG_TYPE_LANGUAGE) pkg = re.match(pattern, line) if comment is not None: continue if pkg is not None: pkg_type = pkg.group(1) pkg_name = pkg.group(2) if not pkg_type in pkgs: pkgs[pkg_type] = [pkg_name] else: pkgs[pkg_type].append(pkg_name) return pkgs ''' This following function parses a full manifest and return a list object with packages. ''' def parse_full_manifest(self): installed_pkgs = list() if not os.path.exists(self.full_manifest): bb.note('full manifest not exist') return installed_pkgs with open(self.full_manifest, 'r') as manifest: for pkg in manifest.read().split('\n'): installed_pkgs.append(pkg.strip()) return installed_pkgs class RpmManifest(Manifest): """ Returns a dictionary object with mip and mlp packages. """ def _split_multilib(self, pkg_list): pkgs = dict() for pkg in pkg_list.split(): pkg_type = self.PKG_TYPE_MUST_INSTALL ml_variants = self.d.getVar('MULTILIB_VARIANTS', True).split() for ml_variant in ml_variants: if pkg.startswith(ml_variant + '-'): pkg_type = self.PKG_TYPE_MULTILIB if not pkg_type in pkgs: pkgs[pkg_type] = pkg else: pkgs[pkg_type] += " " + pkg return pkgs def create_initial(self): pkgs = dict() with open(self.initial_manifest, "w+") as manifest: manifest.write(self.initial_manifest_file_header) for var in self.var_maps[self.manifest_type]: if var in self.vars_to_split: split_pkgs = self._split_multilib(self.d.getVar(var, True)) if split_pkgs is not None: pkgs = dict(pkgs.items() + split_pkgs.items()) else: pkg_list = self.d.getVar(var, True) if pkg_list is not None: pkgs[self.var_maps[self.manifest_type][var]] = self.d.getVar(var, True) for pkg_type in pkgs: for pkg in pkgs[pkg_type].split(): manifest.write("%s,%s\n" % (pkg_type, pkg)) def create_final(self): pass def create_full(self, pm): pass class OpkgManifest(Manifest): """ Returns a dictionary object with mip and mlp packages. """ def _split_multilib(self, pkg_list): pkgs = dict() for pkg in pkg_list.split(): pkg_type = self.PKG_TYPE_MUST_INSTALL ml_variants = self.d.getVar('MULTILIB_VARIANTS', True).split() for ml_variant in ml_variants: if pkg.startswith(ml_variant + '-'): pkg_type = self.PKG_TYPE_MULTILIB if not pkg_type in pkgs: pkgs[pkg_type] = pkg else: pkgs[pkg_type] += " " + pkg return pkgs def create_initial(self): pkgs = dict() with open(self.initial_manifest, "w+") as manifest: manifest.write(self.initial_manifest_file_header) for var in self.var_maps[self.manifest_type]: if var in self.vars_to_split: split_pkgs = self._split_multilib(self.d.getVar(var, True)) if split_pkgs is not None: pkgs = dict(pkgs.items() + split_pkgs.items()) else: pkg_list = self.d.getVar(var, True) if pkg_list is not None: pkgs[self.var_maps[self.manifest_type][var]] = self.d.getVar(var, True) for pkg_type in pkgs: for pkg in pkgs[pkg_type].split(): manifest.write("%s,%s\n" % (pkg_type, pkg)) def create_final(self): pass def create_full(self, pm): if not os.path.exists(self.initial_manifest): self.create_initial() initial_manifest = self.parse_initial_manifest() pkgs_to_install = list() for pkg_type in initial_manifest: pkgs_to_install += initial_manifest[pkg_type] if len(pkgs_to_install) == 0: return output = pm.dummy_install(pkgs_to_install) with open(self.full_manifest, 'w+') as manifest: pkg_re = re.compile('^Installing ([^ ]+) [^ ].*') for line in set(output.split('\n')): m = pkg_re.match(line) if m: manifest.write(m.group(1) + '\n') return class DpkgManifest(Manifest): def create_initial(self): with open(self.initial_manifest, "w+") as manifest: manifest.write(self.initial_manifest_file_header) for var in self.var_maps[self.manifest_type]: pkg_list = self.d.getVar(var, True) if pkg_list is None: continue for pkg in pkg_list.split(): manifest.write("%s,%s\n" % (self.var_maps[self.manifest_type][var], pkg)) def create_final(self): pass def create_full(self, pm): pass def create_manifest(d, final_manifest=False, manifest_dir=None, manifest_type=Manifest.MANIFEST_TYPE_IMAGE): manifest_map = {'rpm': RpmManifest, 'ipk': OpkgManifest, 'deb': DpkgManifest} manifest = manifest_map[d.getVar('IMAGE_PKGTYPE', True)](d, manifest_dir, manifest_type) if final_manifest: manifest.create_final() else: manifest.create_initial() if __name__ == "__main__": pass ########NEW FILE######## __FILENAME__ = package def runstrip(arg): # Function to strip a single file, called from split_and_strip_files below # A working 'file' (one which works on the target architecture) # # The elftype is a bit pattern (explained in split_and_strip_files) to tell # us what type of file we're processing... # 4 - executable # 8 - shared library # 16 - kernel module import commands, stat, subprocess (file, elftype, strip) = arg newmode = None if not os.access(file, os.W_OK) or os.access(file, os.R_OK): origmode = os.stat(file)[stat.ST_MODE] newmode = origmode | stat.S_IWRITE | stat.S_IREAD os.chmod(file, newmode) extraflags = "" # kernel module if elftype & 16: extraflags = "--strip-debug --remove-section=.comment --remove-section=.note --preserve-dates" # .so and shared library elif ".so" in file and elftype & 8: extraflags = "--remove-section=.comment --remove-section=.note --strip-unneeded" # shared or executable: elif elftype & 8 or elftype & 4: extraflags = "--remove-section=.comment --remove-section=.note" stripcmd = "'%s' %s '%s'" % (strip, extraflags, file) bb.debug(1, "runstrip: %s" % stripcmd) ret = subprocess.call(stripcmd, shell=True) if newmode: os.chmod(file, origmode) if ret: bb.error("runstrip: '%s' strip command failed" % stripcmd) return def file_translate(file): ft = file.replace("@", "@at@") ft = ft.replace(" ", "@space@") ft = ft.replace("\t", "@tab@") ft = ft.replace("[", "@openbrace@") ft = ft.replace("]", "@closebrace@") ft = ft.replace("_", "@underscore@") return ft def filedeprunner(arg): import re, subprocess, shlex (pkg, pkgfiles, rpmdeps, pkgdest) = arg provides = {} requires = {} r = re.compile(r'[<>=]+ +[^ ]*') def process_deps(pipe, pkg, pkgdest, provides, requires): for line in pipe: f = line.split(" ", 1)[0].strip() line = line.split(" ", 1)[1].strip() if line.startswith("Requires:"): i = requires elif line.startswith("Provides:"): i = provides else: continue file = f.replace(pkgdest + "/" + pkg, "") file = file_translate(file) value = line.split(":", 1)[1].strip() value = r.sub(r'(\g<0>)', value) if value.startswith("rpmlib("): continue if value == "python": continue if file not in i: i[file] = [] i[file].append(value) return provides, requires try: dep_popen = subprocess.Popen(shlex.split(rpmdeps) + pkgfiles, stdout=subprocess.PIPE) provides, requires = process_deps(dep_popen.stdout, pkg, pkgdest, provides, requires) except OSError as e: bb.error("rpmdeps: '%s' command failed, '%s'" % (shlex.split(rpmdeps) + pkgfiles, e)) raise e return (pkg, provides, requires) ########NEW FILE######## __FILENAME__ = packagedata import codecs def packaged(pkg, d): return os.access(get_subpkgedata_fn(pkg, d) + '.packaged', os.R_OK) def read_pkgdatafile(fn): pkgdata = {} def decode(str): c = codecs.getdecoder("string_escape") return c(str)[0] if os.access(fn, os.R_OK): import re f = open(fn, 'r') lines = f.readlines() f.close() r = re.compile("([^:]+):\s*(.*)") for l in lines: m = r.match(l) if m: pkgdata[m.group(1)] = decode(m.group(2)) return pkgdata def get_subpkgedata_fn(pkg, d): return d.expand('${PKGDATA_DIR}/runtime/%s' % pkg) def has_subpkgdata(pkg, d): return os.access(get_subpkgedata_fn(pkg, d), os.R_OK) def read_subpkgdata(pkg, d): return read_pkgdatafile(get_subpkgedata_fn(pkg, d)) def has_pkgdata(pn, d): fn = d.expand('${PKGDATA_DIR}/%s' % pn) return os.access(fn, os.R_OK) def read_pkgdata(pn, d): fn = d.expand('${PKGDATA_DIR}/%s' % pn) return read_pkgdatafile(fn) # # Collapse FOO_pkg variables into FOO # def read_subpkgdata_dict(pkg, d): ret = {} subd = read_pkgdatafile(get_subpkgedata_fn(pkg, d)) for var in subd: newvar = var.replace("_" + pkg, "") if newvar == var and var + "_" + pkg in subd: continue ret[newvar] = subd[var] return ret def _pkgmap(d): """Return a dictionary mapping package to recipe name.""" pkgdatadir = d.getVar("PKGDATA_DIR", True) pkgmap = {} try: files = os.listdir(pkgdatadir) except OSError: bb.warn("No files in %s?" % pkgdatadir) files = [] for pn in filter(lambda f: not os.path.isdir(os.path.join(pkgdatadir, f)), files): try: pkgdata = read_pkgdatafile(os.path.join(pkgdatadir, pn)) except OSError: continue packages = pkgdata.get("PACKAGES") or "" for pkg in packages.split(): pkgmap[pkg] = pn return pkgmap def pkgmap(d): """Return a dictionary mapping package to recipe name. Cache the mapping in the metadata""" pkgmap_data = d.getVar("__pkgmap_data", False) if pkgmap_data is None: pkgmap_data = _pkgmap(d) d.setVar("__pkgmap_data", pkgmap_data) return pkgmap_data def recipename(pkg, d): """Return the recipe name for the given binary package name.""" return pkgmap(d).get(pkg) ########NEW FILE######## __FILENAME__ = packagegroup import itertools def is_optional(feature, d): packages = d.getVar("FEATURE_PACKAGES_%s" % feature, True) if packages: return bool(d.getVarFlag("FEATURE_PACKAGES_%s" % feature, "optional")) else: return bool(d.getVarFlag("PACKAGE_GROUP_%s" % feature, "optional")) def packages(features, d): for feature in features: packages = d.getVar("FEATURE_PACKAGES_%s" % feature, True) if not packages: packages = d.getVar("PACKAGE_GROUP_%s" % feature, True) for pkg in (packages or "").split(): yield pkg def required_packages(features, d): req = filter(lambda feature: not is_optional(feature, d), features) return packages(req, d) def optional_packages(features, d): opt = filter(lambda feature: is_optional(feature, d), features) return packages(opt, d) def active_packages(features, d): return itertools.chain(required_packages(features, d), optional_packages(features, d)) def active_recipes(features, d): import oe.packagedata for pkg in active_packages(features, d): recipe = oe.packagedata.recipename(pkg, d) if recipe: yield recipe ########NEW FILE######## __FILENAME__ = package_manager from abc import ABCMeta, abstractmethod import os import glob import subprocess import shutil import multiprocessing import re import bb import tempfile # this can be used by all PM backends to create the index files in parallel def create_index(arg): index_cmd = arg try: bb.note("Executing '%s' ..." % index_cmd) subprocess.check_output(index_cmd, stderr=subprocess.STDOUT, shell=True) except subprocess.CalledProcessError as e: return("Index creation command '%s' failed with return code %d:\n%s" % (e.cmd, e.returncode, e.output)) return None class Indexer(object): __metaclass__ = ABCMeta def __init__(self, d, deploy_dir): self.d = d self.deploy_dir = deploy_dir @abstractmethod def write_index(self): pass class RpmIndexer(Indexer): def get_ml_prefix_and_os_list(self, arch_var=None, os_var=None): package_archs = { 'default': [], } target_os = { 'default': "", } if arch_var is not None and os_var is not None: package_archs['default'] = self.d.getVar(arch_var, True).split() package_archs['default'].reverse() target_os['default'] = self.d.getVar(os_var, True).strip() else: package_archs['default'] = self.d.getVar("PACKAGE_ARCHS", True).split() # arch order is reversed. This ensures the -best- match is # listed first! package_archs['default'].reverse() target_os['default'] = self.d.getVar("TARGET_OS", True).strip() multilibs = self.d.getVar('MULTILIBS', True) or "" for ext in multilibs.split(): eext = ext.split(':') if len(eext) > 1 and eext[0] == 'multilib': localdata = bb.data.createCopy(self.d) default_tune_key = "DEFAULTTUNE_virtclass-multilib-" + eext[1] default_tune = localdata.getVar(default_tune_key, False) if default_tune: localdata.setVar("DEFAULTTUNE", default_tune) bb.data.update_data(localdata) package_archs[eext[1]] = localdata.getVar('PACKAGE_ARCHS', True).split() package_archs[eext[1]].reverse() target_os[eext[1]] = localdata.getVar("TARGET_OS", True).strip() ml_prefix_list = dict() for mlib in package_archs: if mlib == 'default': ml_prefix_list[mlib] = package_archs[mlib] else: ml_prefix_list[mlib] = list() for arch in package_archs[mlib]: if arch in ['all', 'noarch', 'any']: ml_prefix_list[mlib].append(arch) else: ml_prefix_list[mlib].append(mlib + "_" + arch) return (ml_prefix_list, target_os) def write_index(self): sdk_pkg_archs = (self.d.getVar('SDK_PACKAGE_ARCHS', True) or "").replace('-', '_').split() all_mlb_pkg_archs = (self.d.getVar('ALL_MULTILIB_PACKAGE_ARCHS', True) or "").replace('-', '_').split() mlb_prefix_list = self.get_ml_prefix_and_os_list()[0] archs = set() for item in mlb_prefix_list: archs = archs.union(set(i.replace('-', '_') for i in mlb_prefix_list[item])) if len(archs) == 0: archs = archs.union(set(all_mlb_pkg_archs)) archs = archs.union(set(sdk_pkg_archs)) rpm_createrepo = bb.utils.which(os.getenv('PATH'), "createrepo") index_cmds = [] rpm_dirs_found = False for arch in archs: arch_dir = os.path.join(self.deploy_dir, arch) if not os.path.isdir(arch_dir): continue index_cmds.append("%s --update -q %s" % (rpm_createrepo, arch_dir)) rpm_dirs_found = True if not rpm_dirs_found: bb.note("There are no packages in %s" % self.deploy_dir) return nproc = multiprocessing.cpu_count() pool = bb.utils.multiprocessingpool(nproc) results = list(pool.imap(create_index, index_cmds)) pool.close() pool.join() for result in results: if result is not None: return(result) class OpkgIndexer(Indexer): def write_index(self): arch_vars = ["ALL_MULTILIB_PACKAGE_ARCHS", "SDK_PACKAGE_ARCHS", "MULTILIB_ARCHS"] opkg_index_cmd = bb.utils.which(os.getenv('PATH'), "opkg-make-index") if not os.path.exists(os.path.join(self.deploy_dir, "Packages")): open(os.path.join(self.deploy_dir, "Packages"), "w").close() index_cmds = [] for arch_var in arch_vars: archs = self.d.getVar(arch_var, True) if archs is None: continue for arch in archs.split(): pkgs_dir = os.path.join(self.deploy_dir, arch) pkgs_file = os.path.join(pkgs_dir, "Packages") if not os.path.isdir(pkgs_dir): continue if not os.path.exists(pkgs_file): open(pkgs_file, "w").close() index_cmds.append('%s -r %s -p %s -m %s' % (opkg_index_cmd, pkgs_file, pkgs_file, pkgs_dir)) if len(index_cmds) == 0: bb.note("There are no packages in %s!" % self.deploy_dir) return nproc = multiprocessing.cpu_count() pool = bb.utils.multiprocessingpool(nproc) results = list(pool.imap(create_index, index_cmds)) pool.close() pool.join() for result in results: if result is not None: return(result) class DpkgIndexer(Indexer): def write_index(self): pkg_archs = self.d.getVar('PACKAGE_ARCHS', True) if pkg_archs is not None: arch_list = pkg_archs.split() sdk_pkg_archs = self.d.getVar('SDK_PACKAGE_ARCHS', True) if sdk_pkg_archs is not None: for a in sdk_pkg_archs.split(): if a not in pkg_archs: arch_list.append(a) apt_ftparchive = bb.utils.which(os.getenv('PATH'), "apt-ftparchive") gzip = bb.utils.which(os.getenv('PATH'), "gzip") index_cmds = [] deb_dirs_found = False for arch in arch_list: arch_dir = os.path.join(self.deploy_dir, arch) if not os.path.isdir(arch_dir): continue cmd = "cd %s; PSEUDO_UNLOAD=1 %s packages . > Packages;" % (arch_dir, apt_ftparchive) cmd += "%s -fc Packages > Packages.gz;" % gzip with open(os.path.join(arch_dir, "Release"), "w+") as release: release.write("Label: %s\n" % arch) cmd += "PSEUDO_UNLOAD=1 %s release . >> Release" % apt_ftparchive index_cmds.append(cmd) deb_dirs_found = True if not deb_dirs_found: bb.note("There are no packages in %s" % self.deploy_dir) return nproc = multiprocessing.cpu_count() pool = bb.utils.multiprocessingpool(nproc) results = list(pool.imap(create_index, index_cmds)) pool.close() pool.join() for result in results: if result is not None: return(result) class PkgsList(object): __metaclass__ = ABCMeta def __init__(self, d, rootfs_dir): self.d = d self.rootfs_dir = rootfs_dir @abstractmethod def list(self, format=None): pass class RpmPkgsList(PkgsList): def __init__(self, d, rootfs_dir, arch_var=None, os_var=None): super(RpmPkgsList, self).__init__(d, rootfs_dir) self.rpm_cmd = bb.utils.which(os.getenv('PATH'), "rpm") self.image_rpmlib = os.path.join(self.rootfs_dir, 'var/lib/rpm') self.ml_prefix_list, self.ml_os_list = \ RpmIndexer(d, rootfs_dir).get_ml_prefix_and_os_list(arch_var, os_var) ''' Translate the RPM/Smart format names to the OE multilib format names ''' def _pkg_translate_smart_to_oe(self, pkg, arch): new_pkg = pkg fixed_arch = arch.replace('_', '-') found = 0 for mlib in self.ml_prefix_list: for cmp_arch in self.ml_prefix_list[mlib]: fixed_cmp_arch = cmp_arch.replace('_', '-') if fixed_arch == fixed_cmp_arch: if mlib == 'default': new_pkg = pkg new_arch = cmp_arch else: new_pkg = mlib + '-' + pkg # We need to strip off the ${mlib}_ prefix on the arch new_arch = cmp_arch.replace(mlib + '_', '') # Workaround for bug 3565. Simply look to see if we # know of a package with that name, if not try again! filename = os.path.join(self.d.getVar('PKGDATA_DIR', True), 'runtime-reverse', new_pkg) if os.path.exists(filename): found = 1 break if found == 1 and fixed_arch == fixed_cmp_arch: break #bb.note('%s, %s -> %s, %s' % (pkg, arch, new_pkg, new_arch)) return new_pkg, new_arch def _list_pkg_deps(self): cmd = [bb.utils.which(os.getenv('PATH'), "rpmresolve"), "-t", self.image_rpmlib] try: output = subprocess.check_output(cmd, stderr=subprocess.STDOUT).strip() except subprocess.CalledProcessError as e: bb.fatal("Cannot get the package dependencies. Command '%s' " "returned %d:\n%s" % (' '.join(cmd), e.returncode, e.output)) return output def list(self, format=None): if format == "deps": return self._list_pkg_deps() cmd = self.rpm_cmd + ' --root ' + self.rootfs_dir cmd += ' -D "_dbpath /var/lib/rpm" -qa' cmd += " --qf '[%{NAME} %{ARCH} %{VERSION} %{PACKAGEORIGIN}\n]'" try: # bb.note(cmd) tmp_output = subprocess.check_output(cmd, stderr=subprocess.STDOUT, shell=True).strip() except subprocess.CalledProcessError as e: bb.fatal("Cannot get the installed packages list. Command '%s' " "returned %d:\n%s" % (cmd, e.returncode, e.output)) output = list() for line in tmp_output.split('\n'): if len(line.strip()) == 0: continue pkg = line.split()[0] arch = line.split()[1] ver = line.split()[2] pkgorigin = line.split()[3] new_pkg, new_arch = self._pkg_translate_smart_to_oe(pkg, arch) if format == "arch": output.append('%s %s' % (new_pkg, new_arch)) elif format == "file": output.append('%s %s %s' % (new_pkg, pkgorigin, new_arch)) elif format == "ver": output.append('%s %s %s' % (new_pkg, new_arch, ver)) else: output.append('%s' % (new_pkg)) output.sort() return '\n'.join(output) class OpkgPkgsList(PkgsList): def __init__(self, d, rootfs_dir, config_file): super(OpkgPkgsList, self).__init__(d, rootfs_dir) self.opkg_cmd = bb.utils.which(os.getenv('PATH'), "opkg-cl") self.opkg_args = "-f %s -o %s " % (config_file, rootfs_dir) self.opkg_args += self.d.getVar("OPKG_ARGS", True) def list(self, format=None): opkg_query_cmd = bb.utils.which(os.getenv('PATH'), "opkg-query-helper.py") if format == "arch": cmd = "%s %s status | %s -a" % \ (self.opkg_cmd, self.opkg_args, opkg_query_cmd) elif format == "file": cmd = "%s %s status | %s -f" % \ (self.opkg_cmd, self.opkg_args, opkg_query_cmd) elif format == "ver": cmd = "%s %s status | %s -v" % \ (self.opkg_cmd, self.opkg_args, opkg_query_cmd) elif format == "deps": cmd = "%s %s status | %s" % \ (self.opkg_cmd, self.opkg_args, opkg_query_cmd) else: cmd = "%s %s list_installed | cut -d' ' -f1" % \ (self.opkg_cmd, self.opkg_args) try: output = subprocess.check_output(cmd, stderr=subprocess.STDOUT, shell=True).strip() except subprocess.CalledProcessError as e: bb.fatal("Cannot get the installed packages list. Command '%s' " "returned %d:\n%s" % (cmd, e.returncode, e.output)) if output and format == "file": tmp_output = "" for line in output.split('\n'): pkg, pkg_file, pkg_arch = line.split() full_path = os.path.join(self.rootfs_dir, pkg_arch, pkg_file) if os.path.exists(full_path): tmp_output += "%s %s %s\n" % (pkg, full_path, pkg_arch) else: tmp_output += "%s %s %s\n" % (pkg, pkg_file, pkg_arch) output = tmp_output return output class DpkgPkgsList(PkgsList): def list(self, format=None): cmd = [bb.utils.which(os.getenv('PATH'), "dpkg-query"), "--admindir=%s/var/lib/dpkg" % self.rootfs_dir, "-W"] if format == "arch": cmd.append("-f=${Package} ${PackageArch}\n") elif format == "file": cmd.append("-f=${Package} ${Package}_${Version}_${Architecture}.deb ${PackageArch}\n") elif format == "ver": cmd.append("-f=${Package} ${PackageArch} ${Version}\n") elif format == "deps": cmd.append("-f=Package: ${Package}\nDepends: ${Depends}\nRecommends: ${Recommends}\n\n") else: cmd.append("-f=${Package}\n") try: output = subprocess.check_output(cmd, stderr=subprocess.STDOUT).strip() except subprocess.CalledProcessError as e: bb.fatal("Cannot get the installed packages list. Command '%s' " "returned %d:\n%s" % (' '.join(cmd), e.returncode, e.output)) if format == "file": tmp_output = "" for line in tuple(output.split('\n')): pkg, pkg_file, pkg_arch = line.split() full_path = os.path.join(self.rootfs_dir, pkg_arch, pkg_file) if os.path.exists(full_path): tmp_output += "%s %s %s\n" % (pkg, full_path, pkg_arch) else: tmp_output += "%s %s %s\n" % (pkg, pkg_file, pkg_arch) output = tmp_output elif format == "deps": opkg_query_cmd = bb.utils.which(os.getenv('PATH'), "opkg-query-helper.py") file_out = tempfile.NamedTemporaryFile() file_out.write(output) file_out.flush() try: output = subprocess.check_output("cat %s | %s" % (file_out.name, opkg_query_cmd), stderr=subprocess.STDOUT, shell=True) except subprocess.CalledProcessError as e: file_out.close() bb.fatal("Cannot compute packages dependencies. Command '%s' " "returned %d:\n%s" % (e.cmd, e.returncode, e.output)) file_out.close() return output class PackageManager(object): """ This is an abstract class. Do not instantiate this directly. """ __metaclass__ = ABCMeta def __init__(self, d): self.d = d self.deploy_dir = None self.deploy_lock = None self.feed_uris = self.d.getVar('PACKAGE_FEED_URIS', True) or "" """ Update the package manager package database. """ @abstractmethod def update(self): pass """ Install a list of packages. 'pkgs' is a list object. If 'attempt_only' is True, installation failures are ignored. """ @abstractmethod def install(self, pkgs, attempt_only=False): pass """ Remove a list of packages. 'pkgs' is a list object. If 'with_dependencies' is False, the any dependencies are left in place. """ @abstractmethod def remove(self, pkgs, with_dependencies=True): pass """ This function creates the index files """ @abstractmethod def write_index(self): pass @abstractmethod def remove_packaging_data(self): pass @abstractmethod def list_installed(self, format=None): pass @abstractmethod def insert_feeds_uris(self): pass """ Install complementary packages based upon the list of currently installed packages e.g. locales, *-dev, *-dbg, etc. This will only attempt to install these packages, if they don't exist then no error will occur. Note: every backend needs to call this function explicitly after the normal package installation """ def install_complementary(self, globs=None): # we need to write the list of installed packages to a file because the # oe-pkgdata-util reads it from a file installed_pkgs_file = os.path.join(self.d.getVar('WORKDIR', True), "installed_pkgs.txt") with open(installed_pkgs_file, "w+") as installed_pkgs: installed_pkgs.write(self.list_installed("arch")) if globs is None: globs = self.d.getVar('IMAGE_INSTALL_COMPLEMENTARY', True) split_linguas = set() for translation in self.d.getVar('IMAGE_LINGUAS', True).split(): split_linguas.add(translation) split_linguas.add(translation.split('-')[0]) split_linguas = sorted(split_linguas) for lang in split_linguas: globs += " *-locale-%s" % lang if globs is None: return cmd = [bb.utils.which(os.getenv('PATH'), "oe-pkgdata-util"), "glob", self.d.getVar('PKGDATA_DIR', True), installed_pkgs_file, globs] try: bb.note("Installing complementary packages ...") complementary_pkgs = subprocess.check_output(cmd, stderr=subprocess.STDOUT) except subprocess.CalledProcessError as e: bb.fatal("Could not compute complementary packages list. Command " "'%s' returned %d:\n%s" % (' '.join(cmd), e.returncode, e.output)) self.install(complementary_pkgs.split(), attempt_only=True) def deploy_dir_lock(self): if self.deploy_dir is None: raise RuntimeError("deploy_dir is not set!") lock_file_name = os.path.join(self.deploy_dir, "deploy.lock") self.deploy_lock = bb.utils.lockfile(lock_file_name) def deploy_dir_unlock(self): if self.deploy_lock is None: return bb.utils.unlockfile(self.deploy_lock) self.deploy_lock = None class RpmPM(PackageManager): def __init__(self, d, target_rootfs, target_vendor, task_name='target', providename=None, arch_var=None, os_var=None): super(RpmPM, self).__init__(d) self.target_rootfs = target_rootfs self.target_vendor = target_vendor self.task_name = task_name self.providename = providename self.fullpkglist = list() self.deploy_dir = self.d.getVar('DEPLOY_DIR_RPM', True) self.etcrpm_dir = os.path.join(self.target_rootfs, "etc/rpm") self.install_dir = os.path.join(self.target_rootfs, "install") self.rpm_cmd = bb.utils.which(os.getenv('PATH'), "rpm") self.smart_cmd = bb.utils.which(os.getenv('PATH'), "smart") self.smart_opt = "--data-dir=" + os.path.join(target_rootfs, 'var/lib/smart') self.scriptlet_wrapper = self.d.expand('${WORKDIR}/scriptlet_wrapper') self.solution_manifest = self.d.expand('${T}/saved/%s_solution' % self.task_name) self.saved_rpmlib = self.d.expand('${T}/saved/%s' % self.task_name) self.image_rpmlib = os.path.join(self.target_rootfs, 'var/lib/rpm') if not os.path.exists(self.d.expand('${T}/saved')): bb.utils.mkdirhier(self.d.expand('${T}/saved')) self.indexer = RpmIndexer(self.d, self.deploy_dir) self.pkgs_list = RpmPkgsList(self.d, self.target_rootfs, arch_var, os_var) self.ml_prefix_list, self.ml_os_list = self.indexer.get_ml_prefix_and_os_list(arch_var, os_var) def insert_feeds_uris(self): if self.feed_uris == "": return # List must be prefered to least preferred order default_platform_extra = set() platform_extra = set() bbextendvariant = self.d.getVar('BBEXTENDVARIANT', True) or "" for mlib in self.ml_os_list: for arch in self.ml_prefix_list[mlib]: plt = arch.replace('-', '_') + '-.*-' + self.ml_os_list[mlib] if mlib == bbextendvariant: default_platform_extra.add(plt) else: platform_extra.add(plt) platform_extra = platform_extra.union(default_platform_extra) arch_list = [] for canonical_arch in platform_extra: arch = canonical_arch.split('-')[0] if not os.path.exists(os.path.join(self.deploy_dir, arch)): continue arch_list.append(arch) uri_iterator = 0 channel_priority = 10 + 5 * len(self.feed_uris.split()) * len(arch_list) for uri in self.feed_uris.split(): for arch in arch_list: bb.note('Note: adding Smart channel url%d%s (%s)' % (uri_iterator, arch, channel_priority)) self._invoke_smart('channel --add url%d-%s type=rpm-md baseurl=%s/rpm/%s -y' % (uri_iterator, arch, uri, arch)) self._invoke_smart('channel --set url%d-%s priority=%d' % (uri_iterator, arch, channel_priority)) channel_priority -= 5 uri_iterator += 1 ''' Create configs for rpm and smart, and multilib is supported ''' def create_configs(self): target_arch = self.d.getVar('TARGET_ARCH', True) platform = '%s%s-%s' % (target_arch.replace('-', '_'), self.target_vendor, self.ml_os_list['default']) # List must be prefered to least preferred order default_platform_extra = list() platform_extra = list() bbextendvariant = self.d.getVar('BBEXTENDVARIANT', True) or "" for mlib in self.ml_os_list: for arch in self.ml_prefix_list[mlib]: plt = arch.replace('-', '_') + '-.*-' + self.ml_os_list[mlib] if mlib == bbextendvariant: if plt not in default_platform_extra: default_platform_extra.append(plt) else: if plt not in platform_extra: platform_extra.append(plt) platform_extra = default_platform_extra + platform_extra self._create_configs(platform, platform_extra) def _invoke_smart(self, args): cmd = "%s %s %s" % (self.smart_cmd, self.smart_opt, args) # bb.note(cmd) try: complementary_pkgs = subprocess.check_output(cmd, stderr=subprocess.STDOUT, shell=True) # bb.note(complementary_pkgs) return complementary_pkgs except subprocess.CalledProcessError as e: bb.fatal("Could not invoke smart. Command " "'%s' returned %d:\n%s" % (cmd, e.returncode, e.output)) def _search_pkg_name_in_feeds(self, pkg, feed_archs): for arch in feed_archs: arch = arch.replace('-', '_') for p in self.fullpkglist: regex_match = r"^%s-[^-]*-[^-]*@%s$" % \ (re.escape(pkg), re.escape(arch)) if re.match(regex_match, p) is not None: # First found is best match # bb.note('%s -> %s' % (pkg, pkg + '@' + arch)) return pkg + '@' + arch return "" ''' Translate the OE multilib format names to the RPM/Smart format names It searched the RPM/Smart format names in probable multilib feeds first, and then searched the default base feed. ''' def _pkg_translate_oe_to_smart(self, pkgs, attempt_only=False): new_pkgs = list() for pkg in pkgs: new_pkg = pkg # Search new_pkg in probable multilibs first for mlib in self.ml_prefix_list: # Jump the default archs if mlib == 'default': continue subst = pkg.replace(mlib + '-', '') # if the pkg in this multilib feed if subst != pkg: feed_archs = self.ml_prefix_list[mlib] new_pkg = self._search_pkg_name_in_feeds(subst, feed_archs) if not new_pkg: # Failed to translate, package not found! err_msg = '%s not found in the %s feeds (%s).\n' % \ (pkg, mlib, " ".join(feed_archs)) if not attempt_only: err_msg += " ".join(self.fullpkglist) bb.fatal(err_msg) bb.warn(err_msg) else: new_pkgs.append(new_pkg) break # Apparently not a multilib package... if pkg == new_pkg: # Search new_pkg in default archs default_archs = self.ml_prefix_list['default'] new_pkg = self._search_pkg_name_in_feeds(pkg, default_archs) if not new_pkg: err_msg = '%s not found in the base feeds (%s).\n' % \ (pkg, ' '.join(default_archs)) if not attempt_only: err_msg += " ".join(self.fullpkglist) bb.fatal(err_msg) bb.warn(err_msg) else: new_pkgs.append(new_pkg) return new_pkgs def _create_configs(self, platform, platform_extra): # Setup base system configuration bb.note("configuring RPM platform settings") # Configure internal RPM environment when using Smart os.environ['RPM_ETCRPM'] = self.etcrpm_dir bb.utils.mkdirhier(self.etcrpm_dir) # Setup temporary directory -- install... if os.path.exists(self.install_dir): bb.utils.remove(self.install_dir, True) bb.utils.mkdirhier(os.path.join(self.install_dir, 'tmp')) channel_priority = 5 platform_dir = os.path.join(self.etcrpm_dir, "platform") with open(platform_dir, "w+") as platform_fd: platform_fd.write(platform + '\n') for pt in platform_extra: channel_priority += 5 platform_fd.write(re.sub("-linux.*$", "-linux.*\n", pt)) # Tell RPM that the "/" directory exist and is available bb.note("configuring RPM system provides") sysinfo_dir = os.path.join(self.etcrpm_dir, "sysinfo") bb.utils.mkdirhier(sysinfo_dir) with open(os.path.join(sysinfo_dir, "Dirnames"), "w+") as dirnames: dirnames.write("/\n") if self.providename: providename_dir = os.path.join(sysinfo_dir, "Providename") if not os.path.exists(providename_dir): providename_content = '\n'.join(self.providename) providename_content += '\n' open(providename_dir, "w+").write(providename_content) # Configure RPM... we enforce these settings! bb.note("configuring RPM DB settings") # After change the __db.* cache size, log file will not be # generated automatically, that will raise some warnings, # so touch a bare log for rpm write into it. rpmlib_log = os.path.join(self.image_rpmlib, 'log', 'log.0000000001') if not os.path.exists(rpmlib_log): bb.utils.mkdirhier(os.path.join(self.image_rpmlib, 'log')) open(rpmlib_log, 'w+').close() DB_CONFIG_CONTENT = "# ================ Environment\n" \ "set_data_dir .\n" \ "set_create_dir .\n" \ "set_lg_dir ./log\n" \ "set_tmp_dir ./tmp\n" \ "set_flags db_log_autoremove on\n" \ "\n" \ "# -- thread_count must be >= 8\n" \ "set_thread_count 64\n" \ "\n" \ "# ================ Logging\n" \ "\n" \ "# ================ Memory Pool\n" \ "set_cachesize 0 1048576 0\n" \ "set_mp_mmapsize 268435456\n" \ "\n" \ "# ================ Locking\n" \ "set_lk_max_locks 16384\n" \ "set_lk_max_lockers 16384\n" \ "set_lk_max_objects 16384\n" \ "mutex_set_max 163840\n" \ "\n" \ "# ================ Replication\n" db_config_dir = os.path.join(self.image_rpmlib, 'DB_CONFIG') if not os.path.exists(db_config_dir): open(db_config_dir, 'w+').write(DB_CONFIG_CONTENT) # Create database so that smart doesn't complain (lazy init) cmd = "%s --root %s --dbpath /var/lib/rpm -qa > /dev/null" % ( self.rpm_cmd, self.target_rootfs) try: subprocess.check_output(cmd, stderr=subprocess.STDOUT, shell=True) except subprocess.CalledProcessError as e: bb.fatal("Create rpm database failed. Command '%s' " "returned %d:\n%s" % (cmd, e.returncode, e.output)) # Configure smart bb.note("configuring Smart settings") bb.utils.remove(os.path.join(self.target_rootfs, 'var/lib/smart'), True) self._invoke_smart('config --set rpm-root=%s' % self.target_rootfs) self._invoke_smart('config --set rpm-dbpath=/var/lib/rpm') self._invoke_smart('config --set rpm-extra-macros._var=%s' % self.d.getVar('localstatedir', True)) cmd = 'config --set rpm-extra-macros._tmppath=/install/tmp' self._invoke_smart(cmd) # Write common configuration for host and target usage self._invoke_smart('config --set rpm-nolinktos=1') self._invoke_smart('config --set rpm-noparentdirs=1') for i in self.d.getVar('BAD_RECOMMENDATIONS', True).split(): self._invoke_smart('flag --set ignore-recommends %s' % i) # Do the following configurations here, to avoid them being # saved for field upgrade if self.d.getVar('NO_RECOMMENDATIONS', True).strip() == "1": self._invoke_smart('config --set ignore-all-recommends=1') pkg_exclude = self.d.getVar('PACKAGE_EXCLUDE', True) or "" for i in pkg_exclude.split(): self._invoke_smart('flag --set exclude-packages %s' % i) # Optional debugging # self._invoke_smart('config --set rpm-log-level=debug') # cmd = 'config --set rpm-log-file=/tmp/smart-debug-logfile' # self._invoke_smart(cmd) ch_already_added = [] for canonical_arch in platform_extra: arch = canonical_arch.split('-')[0] arch_channel = os.path.join(self.deploy_dir, arch) if os.path.exists(arch_channel) and not arch in ch_already_added: bb.note('Note: adding Smart channel %s (%s)' % (arch, channel_priority)) self._invoke_smart('channel --add %s type=rpm-md baseurl=%s -y' % (arch, arch_channel)) self._invoke_smart('channel --set %s priority=%d' % (arch, channel_priority)) channel_priority -= 5 ch_already_added.append(arch) bb.note('adding Smart RPM DB channel') self._invoke_smart('channel --add rpmsys type=rpm-sys -y') # Construct install scriptlet wrapper. # Scripts need to be ordered when executed, this ensures numeric order. # If we ever run into needing more the 899 scripts, we'll have to. # change num to start with 1000. # SCRIPTLET_FORMAT = "#!/bin/bash\n" \ "\n" \ "export PATH=%s\n" \ "export D=%s\n" \ 'export OFFLINE_ROOT="$D"\n' \ 'export IPKG_OFFLINE_ROOT="$D"\n' \ 'export OPKG_OFFLINE_ROOT="$D"\n' \ "export INTERCEPT_DIR=%s\n" \ "export NATIVE_ROOT=%s\n" \ "\n" \ "$2 $1/$3 $4\n" \ "if [ $? -ne 0 ]; then\n" \ " if [ $4 -eq 1 ]; then\n" \ " mkdir -p $1/etc/rpm-postinsts\n" \ " num=100\n" \ " while [ -e $1/etc/rpm-postinsts/${num}-* ]; do num=$((num + 1)); done\n" \ " name=`head -1 $1/$3 | cut -d\' \' -f 2`\n" \ ' echo "#!$2" > $1/etc/rpm-postinsts/${num}-${name}\n' \ ' echo "# Arg: $4" >> $1/etc/rpm-postinsts/${num}-${name}\n' \ " cat $1/$3 >> $1/etc/rpm-postinsts/${num}-${name}\n" \ " chmod +x $1/etc/rpm-postinsts/${num}-${name}\n" \ " else\n" \ ' echo "Error: pre/post remove scriptlet failed"\n' \ " fi\n" \ "fi\n" intercept_dir = self.d.expand('${WORKDIR}/intercept_scripts') native_root = self.d.getVar('STAGING_DIR_NATIVE', True) scriptlet_content = SCRIPTLET_FORMAT % (os.environ['PATH'], self.target_rootfs, intercept_dir, native_root) open(self.scriptlet_wrapper, 'w+').write(scriptlet_content) bb.note("Note: configuring RPM cross-install scriptlet_wrapper") os.chmod(self.scriptlet_wrapper, 0755) cmd = 'config --set rpm-extra-macros._cross_scriptlet_wrapper=%s' % \ self.scriptlet_wrapper self._invoke_smart(cmd) # Debug to show smart config info # bb.note(self._invoke_smart('config --show')) def update(self): self._invoke_smart('update rpmsys') ''' Install pkgs with smart, the pkg name is oe format ''' def install(self, pkgs, attempt_only=False): bb.note("Installing the following packages: %s" % ' '.join(pkgs)) if attempt_only and len(pkgs) == 0: return pkgs = self._pkg_translate_oe_to_smart(pkgs, attempt_only) if not attempt_only: bb.note('to be installed: %s' % ' '.join(pkgs)) cmd = "%s %s install -y %s" % \ (self.smart_cmd, self.smart_opt, ' '.join(pkgs)) bb.note(cmd) else: bb.note('installing attempt only packages...') bb.note('Attempting %s' % ' '.join(pkgs)) cmd = "%s %s install --attempt -y %s" % \ (self.smart_cmd, self.smart_opt, ' '.join(pkgs)) try: output = subprocess.check_output(cmd.split(), stderr=subprocess.STDOUT) bb.note(output) except subprocess.CalledProcessError as e: bb.fatal("Unable to install packages. Command '%s' " "returned %d:\n%s" % (cmd, e.returncode, e.output)) ''' Remove pkgs with smart, the pkg name is smart/rpm format ''' def remove(self, pkgs, with_dependencies=True): bb.note('to be removed: ' + ' '.join(pkgs)) if not with_dependencies: cmd = "%s -e --nodeps " % self.rpm_cmd cmd += "--root=%s " % self.target_rootfs cmd += "--dbpath=/var/lib/rpm " cmd += "--define='_cross_scriptlet_wrapper %s' " % \ self.scriptlet_wrapper cmd += "--define='_tmppath /install/tmp' %s" % ' '.join(pkgs) else: # for pkg in pkgs: # bb.note('Debug: What required: %s' % pkg) # bb.note(self._invoke_smart('query %s --show-requiredby' % pkg)) cmd = "%s %s remove -y %s" % (self.smart_cmd, self.smart_opt, ' '.join(pkgs)) try: bb.note(cmd) output = subprocess.check_output(cmd, stderr=subprocess.STDOUT, shell=True) bb.note(output) except subprocess.CalledProcessError as e: bb.note("Unable to remove packages. Command '%s' " "returned %d:\n%s" % (cmd, e.returncode, e.output)) def upgrade(self): bb.note('smart upgrade') self._invoke_smart('upgrade') def write_index(self): result = self.indexer.write_index() if result is not None: bb.fatal(result) def remove_packaging_data(self): bb.utils.remove(self.image_rpmlib, True) bb.utils.remove(os.path.join(self.target_rootfs, 'var/lib/smart'), True) bb.utils.remove(os.path.join(self.target_rootfs, 'var/lib/opkg'), True) # remove temp directory bb.utils.remove(self.d.expand('${IMAGE_ROOTFS}/install'), True) def backup_packaging_data(self): # Save the rpmlib for increment rpm image generation if os.path.exists(self.saved_rpmlib): bb.utils.remove(self.saved_rpmlib, True) shutil.copytree(self.image_rpmlib, self.saved_rpmlib, symlinks=True) def recovery_packaging_data(self): # Move the rpmlib back if os.path.exists(self.saved_rpmlib): if os.path.exists(self.image_rpmlib): bb.utils.remove(self.image_rpmlib, True) bb.note('Recovery packaging data') shutil.copytree(self.saved_rpmlib, self.image_rpmlib, symlinks=True) def list_installed(self, format=None): return self.pkgs_list.list(format) ''' If incremental install, we need to determine what we've got, what we need to add, and what to remove... The dump_install_solution will dump and save the new install solution. ''' def dump_install_solution(self, pkgs): bb.note('creating new install solution for incremental install') if len(pkgs) == 0: return pkgs = self._pkg_translate_oe_to_smart(pkgs, False) install_pkgs = list() cmd = "%s %s install -y --dump %s 2>%s" % \ (self.smart_cmd, self.smart_opt, ' '.join(pkgs), self.solution_manifest) try: # Disable rpmsys channel for the fake install self._invoke_smart('channel --disable rpmsys') subprocess.check_output(cmd, stderr=subprocess.STDOUT, shell=True) with open(self.solution_manifest, 'r') as manifest: for pkg in manifest.read().split('\n'): if '@' in pkg: install_pkgs.append(pkg) except subprocess.CalledProcessError as e: bb.note("Unable to dump install packages. Command '%s' " "returned %d:\n%s" % (cmd, e.returncode, e.output)) # Recovery rpmsys channel self._invoke_smart('channel --enable rpmsys') return install_pkgs ''' If incremental install, we need to determine what we've got, what we need to add, and what to remove... The load_old_install_solution will load the previous install solution ''' def load_old_install_solution(self): bb.note('load old install solution for incremental install') installed_pkgs = list() if not os.path.exists(self.solution_manifest): bb.note('old install solution not exist') return installed_pkgs with open(self.solution_manifest, 'r') as manifest: for pkg in manifest.read().split('\n'): if '@' in pkg: installed_pkgs.append(pkg.strip()) return installed_pkgs ''' Dump all available packages in feeds, it should be invoked after the newest rpm index was created ''' def dump_all_available_pkgs(self): available_manifest = self.d.expand('${T}/saved/available_pkgs.txt') available_pkgs = list() cmd = "%s %s query --output %s" % \ (self.smart_cmd, self.smart_opt, available_manifest) try: subprocess.check_output(cmd, stderr=subprocess.STDOUT, shell=True) with open(available_manifest, 'r') as manifest: for pkg in manifest.read().split('\n'): if '@' in pkg: available_pkgs.append(pkg.strip()) except subprocess.CalledProcessError as e: bb.note("Unable to list all available packages. Command '%s' " "returned %d:\n%s" % (cmd, e.returncode, e.output)) self.fullpkglist = available_pkgs return def save_rpmpostinst(self, pkg): mlibs = (self.d.getVar('MULTILIB_GLOBAL_VARIANTS') or "").split() new_pkg = pkg # Remove any multilib prefix from the package name for mlib in mlibs: if mlib in pkg: new_pkg = pkg.replace(mlib + '-', '') break bb.note(' * postponing %s' % new_pkg) saved_dir = self.target_rootfs + self.d.expand('${sysconfdir}/rpm-postinsts/') + new_pkg cmd = self.rpm_cmd + ' -q --scripts --root ' + self.target_rootfs cmd += ' --dbpath=/var/lib/rpm ' + new_pkg cmd += ' | sed -n -e "/^postinstall scriptlet (using .*):$/,/^.* scriptlet (using .*):$/ {/.*/p}"' cmd += ' | sed -e "/postinstall scriptlet (using \(.*\)):$/d"' cmd += ' -e "/^.* scriptlet (using .*):$/d" > %s' % saved_dir try: bb.note(cmd) output = subprocess.check_output(cmd, stderr=subprocess.STDOUT, shell=True).strip() bb.note(output) os.chmod(saved_dir, 0755) except subprocess.CalledProcessError as e: bb.fatal("Invoke save_rpmpostinst failed. Command '%s' " "returned %d:\n%s" % (cmd, e.returncode, e.output)) '''Write common configuration for target usage''' def rpm_setup_smart_target_config(self): bb.utils.remove(os.path.join(self.target_rootfs, 'var/lib/smart'), True) self._invoke_smart('config --set rpm-nolinktos=1') self._invoke_smart('config --set rpm-noparentdirs=1') for i in self.d.getVar('BAD_RECOMMENDATIONS', True).split(): self._invoke_smart('flag --set ignore-recommends %s' % i) self._invoke_smart('channel --add rpmsys type=rpm-sys -y') ''' The rpm db lock files were produced after invoking rpm to query on build system, and they caused the rpm on target didn't work, so we need to unlock the rpm db by removing the lock files. ''' def unlock_rpm_db(self): # Remove rpm db lock files rpm_db_locks = glob.glob('%s/var/lib/rpm/__db.*' % self.target_rootfs) for f in rpm_db_locks: bb.utils.remove(f, True) class OpkgPM(PackageManager): def __init__(self, d, target_rootfs, config_file, archs, task_name='target'): super(OpkgPM, self).__init__(d) self.target_rootfs = target_rootfs self.config_file = config_file self.pkg_archs = archs self.task_name = task_name self.deploy_dir = self.d.getVar("DEPLOY_DIR_IPK", True) self.deploy_lock_file = os.path.join(self.deploy_dir, "deploy.lock") self.opkg_cmd = bb.utils.which(os.getenv('PATH'), "opkg-cl") self.opkg_args = "-f %s -o %s " % (self.config_file, target_rootfs) self.opkg_args += self.d.getVar("OPKG_ARGS", True) opkg_lib_dir = self.d.getVar('OPKGLIBDIR', True) if opkg_lib_dir[0] == "/": opkg_lib_dir = opkg_lib_dir[1:] self.opkg_dir = os.path.join(target_rootfs, opkg_lib_dir, "opkg") bb.utils.mkdirhier(self.opkg_dir) self.saved_opkg_dir = self.d.expand('${T}/saved/%s' % self.task_name) if not os.path.exists(self.d.expand('${T}/saved')): bb.utils.mkdirhier(self.d.expand('${T}/saved')) if (self.d.getVar('BUILD_IMAGES_FROM_FEEDS', True) or "") != "1": self._create_config() else: self._create_custom_config() self.indexer = OpkgIndexer(self.d, self.deploy_dir) """ This function will change a package's status in /var/lib/opkg/status file. If 'packages' is None then the new_status will be applied to all packages """ def mark_packages(self, status_tag, packages=None): status_file = os.path.join(self.opkg_dir, "status") with open(status_file, "r") as sf: with open(status_file + ".tmp", "w+") as tmp_sf: if packages is None: tmp_sf.write(re.sub(r"Package: (.*?)\n((?:[^\n]+\n)*?)Status: (.*)(?:unpacked|installed)", r"Package: \1\n\2Status: \3%s" % status_tag, sf.read())) else: if type(packages).__name__ != "list": raise TypeError("'packages' should be a list object") status = sf.read() for pkg in packages: status = re.sub(r"Package: %s\n((?:[^\n]+\n)*?)Status: (.*)(?:unpacked|installed)" % pkg, r"Package: %s\n\1Status: \2%s" % (pkg, status_tag), status) tmp_sf.write(status) os.rename(status_file + ".tmp", status_file) def _create_custom_config(self): bb.note("Building from feeds activated!") with open(self.config_file, "w+") as config_file: priority = 1 for arch in self.pkg_archs.split(): config_file.write("arch %s %d\n" % (arch, priority)) priority += 5 for line in (self.d.getVar('IPK_FEED_URIS', True) or "").split(): feed_match = re.match("^[ \t]*(.*)##([^ \t]*)[ \t]*$", line) if feed_match is not None: feed_name = feed_match.group(1) feed_uri = feed_match.group(2) bb.note("Add %s feed with URL %s" % (feed_name, feed_uri)) config_file.write("src/gz %s %s\n" % (feed_name, feed_uri)) """ Allow to use package deploy directory contents as quick devel-testing feed. This creates individual feed configs for each arch subdir of those specified as compatible for the current machine. NOTE: Development-helper feature, NOT a full-fledged feed. """ if (self.d.getVar('FEED_DEPLOYDIR_BASE_URI', True) or "") != "": for arch in self.pkg_archs.split(): cfg_file_name = os.path.join(self.target_rootfs, self.d.getVar("sysconfdir", True), "opkg", "local-%s-feed.conf" % arch) with open(cfg_file_name, "w+") as cfg_file: cfg_file.write("src/gz local-%s %s/%s" % arch, self.d.getVar('FEED_DEPLOYDIR_BASE_URI', True), arch) def _create_config(self): with open(self.config_file, "w+") as config_file: priority = 1 for arch in self.pkg_archs.split(): config_file.write("arch %s %d\n" % (arch, priority)) priority += 5 config_file.write("src oe file:%s\n" % self.deploy_dir) for arch in self.pkg_archs.split(): pkgs_dir = os.path.join(self.deploy_dir, arch) if os.path.isdir(pkgs_dir): config_file.write("src oe-%s file:%s\n" % (arch, pkgs_dir)) def insert_feeds_uris(self): if self.feed_uris == "": return rootfs_config = os.path.join('%s/etc/opkg/base-feeds.conf' % self.target_rootfs) with open(rootfs_config, "w+") as config_file: uri_iterator = 0 for uri in self.feed_uris.split(): config_file.write("src/gz url-%d %s/ipk\n" % (uri_iterator, uri)) for arch in self.pkg_archs.split(): if not os.path.exists(os.path.join(self.deploy_dir, arch)): continue bb.note('Note: adding opkg channel url-%s-%d (%s)' % (arch, uri_iterator, uri)) config_file.write("src/gz uri-%s-%d %s/ipk/%s\n" % (arch, uri_iterator, uri, arch)) uri_iterator += 1 def update(self): self.deploy_dir_lock() cmd = "%s %s update" % (self.opkg_cmd, self.opkg_args) try: subprocess.check_output(cmd.split(), stderr=subprocess.STDOUT) except subprocess.CalledProcessError as e: self.deploy_dir_unlock() bb.fatal("Unable to update the package index files. Command '%s' " "returned %d:\n%s" % (cmd, e.returncode, e.output)) self.deploy_dir_unlock() def install(self, pkgs, attempt_only=False): if attempt_only and len(pkgs) == 0: return cmd = "%s %s install %s" % (self.opkg_cmd, self.opkg_args, ' '.join(pkgs)) os.environ['D'] = self.target_rootfs os.environ['OFFLINE_ROOT'] = self.target_rootfs os.environ['IPKG_OFFLINE_ROOT'] = self.target_rootfs os.environ['OPKG_OFFLINE_ROOT'] = self.target_rootfs os.environ['INTERCEPT_DIR'] = os.path.join(self.d.getVar('WORKDIR', True), "intercept_scripts") os.environ['NATIVE_ROOT'] = self.d.getVar('STAGING_DIR_NATIVE', True) try: bb.note("Installing the following packages: %s" % ' '.join(pkgs)) bb.note(cmd) output = subprocess.check_output(cmd.split(), stderr=subprocess.STDOUT) bb.note(output) except subprocess.CalledProcessError as e: (bb.fatal, bb.note)[attempt_only]("Unable to install packages. " "Command '%s' returned %d:\n%s" % (cmd, e.returncode, e.output)) def remove(self, pkgs, with_dependencies=True): if with_dependencies: cmd = "%s %s --force-depends --force-remove --force-removal-of-dependent-packages remove %s" % \ (self.opkg_cmd, self.opkg_args, ' '.join(pkgs)) else: cmd = "%s %s --force-depends remove %s" % \ (self.opkg_cmd, self.opkg_args, ' '.join(pkgs)) try: bb.note(cmd) output = subprocess.check_output(cmd.split(), stderr=subprocess.STDOUT) bb.note(output) except subprocess.CalledProcessError as e: bb.fatal("Unable to remove packages. Command '%s' " "returned %d:\n%s" % (e.cmd, e.returncode, e.output)) def write_index(self): self.deploy_dir_lock() result = self.indexer.write_index() self.deploy_dir_unlock() if result is not None: bb.fatal(result) def remove_packaging_data(self): bb.utils.remove(self.opkg_dir, True) # create the directory back, it's needed by PM lock bb.utils.mkdirhier(self.opkg_dir) def list_installed(self, format=None): return OpkgPkgsList(self.d, self.target_rootfs, self.config_file).list(format) def handle_bad_recommendations(self): bad_recommendations = self.d.getVar("BAD_RECOMMENDATIONS", True) or "" if bad_recommendations.strip() == "": return status_file = os.path.join(self.opkg_dir, "status") # If status file existed, it means the bad recommendations has already # been handled if os.path.exists(status_file): return cmd = "%s %s info " % (self.opkg_cmd, self.opkg_args) with open(status_file, "w+") as status: for pkg in bad_recommendations.split(): pkg_info = cmd + pkg try: output = subprocess.check_output(pkg_info.split(), stderr=subprocess.STDOUT).strip() except subprocess.CalledProcessError as e: bb.fatal("Cannot get package info. Command '%s' " "returned %d:\n%s" % (pkg_info, e.returncode, e.output)) if output == "": bb.note("Ignored bad recommendation: '%s' is " "not a package" % pkg) continue for line in output.split('\n'): if line.startswith("Status:"): status.write("Status: deinstall hold not-installed\n") else: status.write(line + "\n") ''' The following function dummy installs pkgs and returns the log of output. ''' def dummy_install(self, pkgs): if len(pkgs) == 0: return # Create an temp dir as opkg root for dummy installation temp_rootfs = self.d.expand('${T}/opkg') temp_opkg_dir = os.path.join(temp_rootfs, 'var/lib/opkg') bb.utils.mkdirhier(temp_opkg_dir) opkg_args = "-f %s -o %s " % (self.config_file, temp_rootfs) opkg_args += self.d.getVar("OPKG_ARGS", True) cmd = "%s %s update" % (self.opkg_cmd, opkg_args) try: subprocess.check_output(cmd, stderr=subprocess.STDOUT, shell=True) except subprocess.CalledProcessError as e: bb.fatal("Unable to update. Command '%s' " "returned %d:\n%s" % (cmd, e.returncode, e.output)) # Dummy installation cmd = "%s %s --noaction install %s " % (self.opkg_cmd, opkg_args, ' '.join(pkgs)) try: output = subprocess.check_output(cmd, stderr=subprocess.STDOUT, shell=True) except subprocess.CalledProcessError as e: bb.fatal("Unable to dummy install packages. Command '%s' " "returned %d:\n%s" % (cmd, e.returncode, e.output)) bb.utils.remove(temp_rootfs, True) return output def backup_packaging_data(self): # Save the opkglib for increment ipk image generation if os.path.exists(self.saved_opkg_dir): bb.utils.remove(self.saved_opkg_dir, True) shutil.copytree(self.opkg_dir, self.saved_opkg_dir, symlinks=True) def recover_packaging_data(self): # Move the opkglib back if os.path.exists(self.saved_opkg_dir): if os.path.exists(self.opkg_dir): bb.utils.remove(self.opkg_dir, True) bb.note('Recover packaging data') shutil.copytree(self.saved_opkg_dir, self.opkg_dir, symlinks=True) class DpkgPM(PackageManager): def __init__(self, d, target_rootfs, archs, base_archs, apt_conf_dir=None): super(DpkgPM, self).__init__(d) self.target_rootfs = target_rootfs self.deploy_dir = self.d.getVar('DEPLOY_DIR_DEB', True) if apt_conf_dir is None: self.apt_conf_dir = self.d.expand("${APTCONF_TARGET}/apt") else: self.apt_conf_dir = apt_conf_dir self.apt_conf_file = os.path.join(self.apt_conf_dir, "apt.conf") self.apt_get_cmd = bb.utils.which(os.getenv('PATH'), "apt-get") self.apt_args = d.getVar("APT_ARGS", True) self._create_configs(archs, base_archs) self.indexer = DpkgIndexer(self.d, self.deploy_dir) """ This function will change a package's status in /var/lib/dpkg/status file. If 'packages' is None then the new_status will be applied to all packages """ def mark_packages(self, status_tag, packages=None): status_file = self.target_rootfs + "/var/lib/dpkg/status" with open(status_file, "r") as sf: with open(status_file + ".tmp", "w+") as tmp_sf: if packages is None: tmp_sf.write(re.sub(r"Package: (.*?)\n((?:[^\n]+\n)*?)Status: (.*)(?:unpacked|installed)", r"Package: \1\n\2Status: \3%s" % status_tag, sf.read())) else: if type(packages).__name__ != "list": raise TypeError("'packages' should be a list object") status = sf.read() for pkg in packages: status = re.sub(r"Package: %s\n((?:[^\n]+\n)*?)Status: (.*)(?:unpacked|installed)" % pkg, r"Package: %s\n\1Status: \2%s" % (pkg, status_tag), status) tmp_sf.write(status) os.rename(status_file + ".tmp", status_file) """ Run the pre/post installs for package "package_name". If package_name is None, then run all pre/post install scriptlets. """ def run_pre_post_installs(self, package_name=None): info_dir = self.target_rootfs + "/var/lib/dpkg/info" suffixes = [(".preinst", "Preinstall"), (".postinst", "Postinstall")] status_file = self.target_rootfs + "/var/lib/dpkg/status" installed_pkgs = [] with open(status_file, "r") as status: for line in status.read().split('\n'): m = re.match("^Package: (.*)", line) if m is not None: installed_pkgs.append(m.group(1)) if package_name is not None and not package_name in installed_pkgs: return os.environ['D'] = self.target_rootfs os.environ['OFFLINE_ROOT'] = self.target_rootfs os.environ['IPKG_OFFLINE_ROOT'] = self.target_rootfs os.environ['OPKG_OFFLINE_ROOT'] = self.target_rootfs os.environ['INTERCEPT_DIR'] = os.path.join(self.d.getVar('WORKDIR', True), "intercept_scripts") os.environ['NATIVE_ROOT'] = self.d.getVar('STAGING_DIR_NATIVE', True) failed_pkgs = [] for pkg_name in installed_pkgs: for suffix in suffixes: p_full = os.path.join(info_dir, pkg_name + suffix[0]) if os.path.exists(p_full): try: bb.note("Executing %s for package: %s ..." % (suffix[1].lower(), pkg_name)) subprocess.check_output(p_full, stderr=subprocess.STDOUT) except subprocess.CalledProcessError as e: bb.note("%s for package %s failed with %d:\n%s" % (suffix[1], pkg_name, e.returncode, e.output)) failed_pkgs.append(pkg_name) break if len(failed_pkgs): self.mark_packages("unpacked", failed_pkgs) def update(self): os.environ['APT_CONFIG'] = self.apt_conf_file self.deploy_dir_lock() cmd = "%s update" % self.apt_get_cmd try: subprocess.check_output(cmd.split(), stderr=subprocess.STDOUT) except subprocess.CalledProcessError as e: bb.fatal("Unable to update the package index files. Command '%s' " "returned %d:\n%s" % (e.cmd, e.returncode, e.output)) self.deploy_dir_unlock() def install(self, pkgs, attempt_only=False): if attempt_only and len(pkgs) == 0: return os.environ['APT_CONFIG'] = self.apt_conf_file cmd = "%s %s install --force-yes --allow-unauthenticated %s" % \ (self.apt_get_cmd, self.apt_args, ' '.join(pkgs)) try: bb.note("Installing the following packages: %s" % ' '.join(pkgs)) subprocess.check_output(cmd.split(), stderr=subprocess.STDOUT) except subprocess.CalledProcessError as e: (bb.fatal, bb.note)[attempt_only]("Unable to install packages. " "Command '%s' returned %d:\n%s" % (cmd, e.returncode, e.output)) # rename *.dpkg-new files/dirs for root, dirs, files in os.walk(self.target_rootfs): for dir in dirs: new_dir = re.sub("\.dpkg-new", "", dir) if dir != new_dir: os.rename(os.path.join(root, dir), os.path.join(root, new_dir)) for file in files: new_file = re.sub("\.dpkg-new", "", file) if file != new_file: os.rename(os.path.join(root, file), os.path.join(root, new_file)) def remove(self, pkgs, with_dependencies=True): if with_dependencies: os.environ['APT_CONFIG'] = self.apt_conf_file cmd = "%s remove %s" % (self.apt_get_cmd, ' '.join(pkgs)) else: cmd = "%s --admindir=%s/var/lib/dpkg --instdir=%s" \ " -r --force-depends %s" % \ (bb.utils.which(os.getenv('PATH'), "dpkg"), self.target_rootfs, self.target_rootfs, ' '.join(pkgs)) try: subprocess.check_output(cmd.split(), stderr=subprocess.STDOUT) except subprocess.CalledProcessError as e: bb.fatal("Unable to remove packages. Command '%s' " "returned %d:\n%s" % (e.cmd, e.returncode, e.output)) def write_index(self): self.deploy_dir_lock() result = self.indexer.write_index() self.deploy_dir_unlock() if result is not None: bb.fatal(result) def insert_feeds_uris(self): if self.feed_uris == "": return sources_conf = os.path.join("%s/etc/apt/sources.list" % self.target_rootfs) arch_list = [] archs = self.d.getVar('PACKAGE_ARCHS', True) for arch in archs.split(): if not os.path.exists(os.path.join(self.deploy_dir, arch)): continue arch_list.append(arch) with open(sources_conf, "w+") as sources_file: for uri in self.feed_uris.split(): for arch in arch_list: bb.note('Note: adding dpkg channel at (%s)' % uri) sources_file.write("deb %s/deb/%s ./\n" % (uri, arch)) def _create_configs(self, archs, base_archs): base_archs = re.sub("_", "-", base_archs) if os.path.exists(self.apt_conf_dir): bb.utils.remove(self.apt_conf_dir, True) bb.utils.mkdirhier(self.apt_conf_dir) bb.utils.mkdirhier(self.apt_conf_dir + "/lists/partial/") bb.utils.mkdirhier(self.apt_conf_dir + "/apt.conf.d/") arch_list = [] for arch in archs.split(): if not os.path.exists(os.path.join(self.deploy_dir, arch)): continue arch_list.append(arch) with open(os.path.join(self.apt_conf_dir, "preferences"), "w+") as prefs_file: priority = 801 for arch in arch_list: prefs_file.write( "Package: *\n" "Pin: release l=%s\n" "Pin-Priority: %d\n\n" % (arch, priority)) priority += 5 pkg_exclude = self.d.getVar('PACKAGE_EXCLUDE', True) or "" for pkg in pkg_exclude: prefs_file.write( "Package: %s\n" "Pin: release *\n" "Pin-Priority: -1\n\n" % pkg) arch_list.reverse() with open(os.path.join(self.apt_conf_dir, "sources.list"), "w+") as sources_file: for arch in arch_list: sources_file.write("deb file:%s/ ./\n" % os.path.join(self.deploy_dir, arch)) with open(self.apt_conf_file, "w+") as apt_conf: with open(self.d.expand("${STAGING_ETCDIR_NATIVE}/apt/apt.conf.sample")) as apt_conf_sample: for line in apt_conf_sample.read().split("\n"): line = re.sub("Architecture \".*\";", "Architecture \"%s\";" % base_archs, line) line = re.sub("#ROOTFS#", self.target_rootfs, line) line = re.sub("#APTCONF#", self.apt_conf_dir, line) apt_conf.write(line + "\n") target_dpkg_dir = "%s/var/lib/dpkg" % self.target_rootfs bb.utils.mkdirhier(os.path.join(target_dpkg_dir, "info")) bb.utils.mkdirhier(os.path.join(target_dpkg_dir, "updates")) if not os.path.exists(os.path.join(target_dpkg_dir, "status")): open(os.path.join(target_dpkg_dir, "status"), "w+").close() if not os.path.exists(os.path.join(target_dpkg_dir, "available")): open(os.path.join(target_dpkg_dir, "available"), "w+").close() def remove_packaging_data(self): bb.utils.remove(os.path.join(self.target_rootfs, self.d.getVar('opkglibdir', True)), True) bb.utils.remove(self.target_rootfs + "/var/lib/dpkg/", True) def fix_broken_dependencies(self): os.environ['APT_CONFIG'] = self.apt_conf_file cmd = "%s %s -f install" % (self.apt_get_cmd, self.apt_args) try: subprocess.check_output(cmd.split(), stderr=subprocess.STDOUT) except subprocess.CalledProcessError as e: bb.fatal("Cannot fix broken dependencies. Command '%s' " "returned %d:\n%s" % (cmd, e.returncode, e.output)) def list_installed(self, format=None): return DpkgPkgsList(self.d, self.target_rootfs).list() def generate_index_files(d): classes = d.getVar('PACKAGE_CLASSES', True).replace("package_", "").split() indexer_map = { "rpm": (RpmIndexer, d.getVar('DEPLOY_DIR_RPM', True)), "ipk": (OpkgIndexer, d.getVar('DEPLOY_DIR_IPK', True)), "deb": (DpkgIndexer, d.getVar('DEPLOY_DIR_DEB', True)) } result = None for pkg_class in classes: if not pkg_class in indexer_map: continue if os.path.exists(indexer_map[pkg_class][1]): result = indexer_map[pkg_class][0](d, indexer_map[pkg_class][1]).write_index() if result is not None: bb.fatal(result) if __name__ == "__main__": """ We should be able to run this as a standalone script, from outside bitbake environment. """ """ TBD """ ########NEW FILE######## __FILENAME__ = patch import oe.path class NotFoundError(bb.BBHandledException): def __init__(self, path): self.path = path def __str__(self): return "Error: %s not found." % self.path class CmdError(bb.BBHandledException): def __init__(self, exitstatus, output): self.status = exitstatus self.output = output def __str__(self): return "Command Error: exit status: %d Output:\n%s" % (self.status, self.output) def runcmd(args, dir = None): import pipes if dir: olddir = os.path.abspath(os.curdir) if not os.path.exists(dir): raise NotFoundError(dir) os.chdir(dir) # print("cwd: %s -> %s" % (olddir, dir)) try: args = [ pipes.quote(str(arg)) for arg in args ] cmd = " ".join(args) # print("cmd: %s" % cmd) (exitstatus, output) = oe.utils.getstatusoutput(cmd) if exitstatus != 0: raise CmdError(exitstatus >> 8, output) return output finally: if dir: os.chdir(olddir) class PatchError(Exception): def __init__(self, msg): self.msg = msg def __str__(self): return "Patch Error: %s" % self.msg class PatchSet(object): defaults = { "strippath": 1 } def __init__(self, dir, d): self.dir = dir self.d = d self.patches = [] self._current = None def current(self): return self._current def Clean(self): """ Clean out the patch set. Generally includes unapplying all patches and wiping out all associated metadata. """ raise NotImplementedError() def Import(self, patch, force): if not patch.get("file"): if not patch.get("remote"): raise PatchError("Patch file must be specified in patch import.") else: patch["file"] = bb.fetch2.localpath(patch["remote"], self.d) for param in PatchSet.defaults: if not patch.get(param): patch[param] = PatchSet.defaults[param] if patch.get("remote"): patch["file"] = bb.data.expand(bb.fetch2.localpath(patch["remote"], self.d), self.d) patch["filemd5"] = bb.utils.md5_file(patch["file"]) def Push(self, force): raise NotImplementedError() def Pop(self, force): raise NotImplementedError() def Refresh(self, remote = None, all = None): raise NotImplementedError() class PatchTree(PatchSet): def __init__(self, dir, d): PatchSet.__init__(self, dir, d) self.patchdir = os.path.join(self.dir, 'patches') self.seriespath = os.path.join(self.dir, 'patches', 'series') bb.utils.mkdirhier(self.patchdir) def _appendPatchFile(self, patch, strippath): with open(self.seriespath, 'a') as f: f.write(os.path.basename(patch) + "," + strippath + "\n") shellcmd = ["cat", patch, ">" , self.patchdir + "/" + os.path.basename(patch)] runcmd(["sh", "-c", " ".join(shellcmd)], self.dir) def _removePatch(self, p): patch = {} patch['file'] = p.split(",")[0] patch['strippath'] = p.split(",")[1] self._applypatch(patch, False, True) def _removePatchFile(self, all = False): if not os.path.exists(self.seriespath): return patches = open(self.seriespath, 'r+').readlines() if all: for p in reversed(patches): self._removePatch(os.path.join(self.patchdir, p.strip())) patches = [] else: self._removePatch(os.path.join(self.patchdir, patches[-1].strip())) patches.pop() with open(self.seriespath, 'w') as f: for p in patches: f.write(p) def Import(self, patch, force = None): """""" PatchSet.Import(self, patch, force) if self._current is not None: i = self._current + 1 else: i = 0 self.patches.insert(i, patch) def _applypatch(self, patch, force = False, reverse = False, run = True): shellcmd = ["cat", patch['file'], "|", "patch", "-p", patch['strippath']] if reverse: shellcmd.append('-R') if not run: return "sh" + "-c" + " ".join(shellcmd) if not force: shellcmd.append('--dry-run') output = runcmd(["sh", "-c", " ".join(shellcmd)], self.dir) if force: return shellcmd.pop(len(shellcmd) - 1) output = runcmd(["sh", "-c", " ".join(shellcmd)], self.dir) if not reverse: self._appendPatchFile(patch['file'], patch['strippath']) return output def Push(self, force = False, all = False, run = True): bb.note("self._current is %s" % self._current) bb.note("patches is %s" % self.patches) if all: for i in self.patches: bb.note("applying patch %s" % i) self._applypatch(i, force) self._current = i else: if self._current is not None: next = self._current + 1 else: next = 0 bb.note("applying patch %s" % self.patches[next]) ret = self._applypatch(self.patches[next], force) self._current = next return ret def Pop(self, force = None, all = None): if all: self._removePatchFile(True) self._current = None else: self._removePatchFile(False) if self._current == 0: self._current = None if self._current is not None: self._current = self._current - 1 def Clean(self): """""" self.Pop(all=True) class GitApplyTree(PatchTree): def __init__(self, dir, d): PatchTree.__init__(self, dir, d) def _applypatch(self, patch, force = False, reverse = False, run = True): def _applypatchhelper(shellcmd, patch, force = False, reverse = False, run = True): if reverse: shellcmd.append('-R') shellcmd.append(patch['file']) if not run: return "sh" + "-c" + " ".join(shellcmd) return runcmd(["sh", "-c", " ".join(shellcmd)], self.dir) try: shellcmd = ["git", "--work-tree=.", "am", "-3", "-p%s" % patch['strippath']] return _applypatchhelper(shellcmd, patch, force, reverse, run) except CmdError: shellcmd = ["git", "--git-dir=.", "apply", "-p%s" % patch['strippath']] return _applypatchhelper(shellcmd, patch, force, reverse, run) class QuiltTree(PatchSet): def _runcmd(self, args, run = True): quiltrc = self.d.getVar('QUILTRCFILE', True) if not run: return ["quilt"] + ["--quiltrc"] + [quiltrc] + args runcmd(["quilt"] + ["--quiltrc"] + [quiltrc] + args, self.dir) def _quiltpatchpath(self, file): return os.path.join(self.dir, "patches", os.path.basename(file)) def __init__(self, dir, d): PatchSet.__init__(self, dir, d) self.initialized = False p = os.path.join(self.dir, 'patches') if not os.path.exists(p): os.makedirs(p) def Clean(self): try: self._runcmd(["pop", "-a", "-f"]) oe.path.remove(os.path.join(self.dir, "patches","series")) except Exception: pass self.initialized = True def InitFromDir(self): # read series -> self.patches seriespath = os.path.join(self.dir, 'patches', 'series') if not os.path.exists(self.dir): raise NotFoundError(self.dir) if os.path.exists(seriespath): series = file(seriespath, 'r') for line in series.readlines(): patch = {} parts = line.strip().split() patch["quiltfile"] = self._quiltpatchpath(parts[0]) patch["quiltfilemd5"] = bb.utils.md5_file(patch["quiltfile"]) if len(parts) > 1: patch["strippath"] = parts[1][2:] self.patches.append(patch) series.close() # determine which patches are applied -> self._current try: output = runcmd(["quilt", "applied"], self.dir) except CmdError: import sys if sys.exc_value.output.strip() == "No patches applied": return else: raise output = [val for val in output.split('\n') if not val.startswith('#')] for patch in self.patches: if os.path.basename(patch["quiltfile"]) == output[-1]: self._current = self.patches.index(patch) self.initialized = True def Import(self, patch, force = None): if not self.initialized: self.InitFromDir() PatchSet.Import(self, patch, force) oe.path.symlink(patch["file"], self._quiltpatchpath(patch["file"]), force=True) f = open(os.path.join(self.dir, "patches","series"), "a"); f.write(os.path.basename(patch["file"]) + " -p" + patch["strippath"]+"\n") f.close() patch["quiltfile"] = self._quiltpatchpath(patch["file"]) patch["quiltfilemd5"] = bb.utils.md5_file(patch["quiltfile"]) # TODO: determine if the file being imported: # 1) is already imported, and is the same # 2) is already imported, but differs self.patches.insert(self._current or 0, patch) def Push(self, force = False, all = False, run = True): # quilt push [-f] args = ["push"] if force: args.append("-f") if all: args.append("-a") if not run: return self._runcmd(args, run) self._runcmd(args) if self._current is not None: self._current = self._current + 1 else: self._current = 0 def Pop(self, force = None, all = None): # quilt pop [-f] args = ["pop"] if force: args.append("-f") if all: args.append("-a") self._runcmd(args) if self._current == 0: self._current = None if self._current is not None: self._current = self._current - 1 def Refresh(self, **kwargs): if kwargs.get("remote"): patch = self.patches[kwargs["patch"]] if not patch: raise PatchError("No patch found at index %s in patchset." % kwargs["patch"]) (type, host, path, user, pswd, parm) = bb.fetch.decodeurl(patch["remote"]) if type == "file": import shutil if not patch.get("file") and patch.get("remote"): patch["file"] = bb.fetch2.localpath(patch["remote"], self.d) shutil.copyfile(patch["quiltfile"], patch["file"]) else: raise PatchError("Unable to do a remote refresh of %s, unsupported remote url scheme %s." % (os.path.basename(patch["quiltfile"]), type)) else: # quilt refresh args = ["refresh"] if kwargs.get("quiltfile"): args.append(os.path.basename(kwargs["quiltfile"])) elif kwargs.get("patch"): args.append(os.path.basename(self.patches[kwargs["patch"]]["quiltfile"])) self._runcmd(args) class Resolver(object): def __init__(self, patchset, terminal): raise NotImplementedError() def Resolve(self): raise NotImplementedError() def Revert(self): raise NotImplementedError() def Finalize(self): raise NotImplementedError() class NOOPResolver(Resolver): def __init__(self, patchset, terminal): self.patchset = patchset self.terminal = terminal def Resolve(self): olddir = os.path.abspath(os.curdir) os.chdir(self.patchset.dir) try: self.patchset.Push() except Exception: import sys os.chdir(olddir) raise # Patch resolver which relies on the user doing all the work involved in the # resolution, with the exception of refreshing the remote copy of the patch # files (the urls). class UserResolver(Resolver): def __init__(self, patchset, terminal): self.patchset = patchset self.terminal = terminal # Force a push in the patchset, then drop to a shell for the user to # resolve any rejected hunks def Resolve(self): olddir = os.path.abspath(os.curdir) os.chdir(self.patchset.dir) try: self.patchset.Push(False) except CmdError as v: # Patch application failed patchcmd = self.patchset.Push(True, False, False) t = self.patchset.d.getVar('T', True) if not t: bb.msg.fatal("Build", "T not set") bb.utils.mkdirhier(t) import random rcfile = "%s/bashrc.%s.%s" % (t, str(os.getpid()), random.random()) f = open(rcfile, "w") f.write("echo '*** Manual patch resolution mode ***'\n") f.write("echo 'Dropping to a shell, so patch rejects can be fixed manually.'\n") f.write("echo 'Run \"quilt refresh\" when patch is corrected, press CTRL+D to exit.'\n") f.write("echo ''\n") f.write(" ".join(patchcmd) + "\n") f.close() os.chmod(rcfile, 0775) self.terminal("bash --rcfile " + rcfile, 'Patch Rejects: Please fix patch rejects manually', self.patchset.d) # Construct a new PatchSet after the user's changes, compare the # sets, checking patches for modifications, and doing a remote # refresh on each. oldpatchset = self.patchset self.patchset = oldpatchset.__class__(self.patchset.dir, self.patchset.d) for patch in self.patchset.patches: oldpatch = None for opatch in oldpatchset.patches: if opatch["quiltfile"] == patch["quiltfile"]: oldpatch = opatch if oldpatch: patch["remote"] = oldpatch["remote"] if patch["quiltfile"] == oldpatch["quiltfile"]: if patch["quiltfilemd5"] != oldpatch["quiltfilemd5"]: bb.note("Patch %s has changed, updating remote url %s" % (os.path.basename(patch["quiltfile"]), patch["remote"])) # user change? remote refresh self.patchset.Refresh(remote=True, patch=self.patchset.patches.index(patch)) else: # User did not fix the problem. Abort. raise PatchError("Patch application failed, and user did not fix and refresh the patch.") except Exception: os.chdir(olddir) raise os.chdir(olddir) ########NEW FILE######## __FILENAME__ = path import errno import glob import shutil import subprocess import os.path def join(*paths): """Like os.path.join but doesn't treat absolute RHS specially""" return os.path.normpath("/".join(paths)) def relative(src, dest): """ Return a relative path from src to dest. >>> relative("/usr/bin", "/tmp/foo/bar") ../../tmp/foo/bar >>> relative("/usr/bin", "/usr/lib") ../lib >>> relative("/tmp", "/tmp/foo/bar") foo/bar """ return os.path.relpath(dest, src) def make_relative_symlink(path): """ Convert an absolute symlink to a relative one """ if not os.path.islink(path): return link = os.readlink(path) if not os.path.isabs(link): return # find the common ancestor directory ancestor = path depth = 0 while ancestor and not link.startswith(ancestor): ancestor = ancestor.rpartition('/')[0] depth += 1 if not ancestor: print("make_relative_symlink() Error: unable to find the common ancestor of %s and its target" % path) return base = link.partition(ancestor)[2].strip('/') while depth > 1: base = "../" + base depth -= 1 os.remove(path) os.symlink(base, path) def format_display(path, metadata): """ Prepare a path for display to the user. """ rel = relative(metadata.getVar("TOPDIR", True), path) if len(rel) > len(path): return path else: return rel def copytree(src, dst): # We could use something like shutil.copytree here but it turns out to # to be slow. It takes twice as long copying to an empty directory. # If dst already has contents performance can be 15 time slower # This way we also preserve hardlinks between files in the tree. bb.utils.mkdirhier(dst) cmd = 'tar -cf - -C %s -p . | tar -xf - -C %s' % (src, dst) check_output(cmd, shell=True, stderr=subprocess.STDOUT) def copyhardlinktree(src, dst): """ Make the hard link when possible, otherwise copy. """ bb.utils.mkdirhier(dst) if os.path.isdir(src) and not len(os.listdir(src)): return if (os.stat(src).st_dev == os.stat(dst).st_dev): # Need to copy directories only with tar first since cp will error if two # writers try and create a directory at the same time cmd = 'cd %s; find . -type d -print | tar -cf - -C %s -p --files-from - --no-recursion | tar -xf - -C %s' % (src, src, dst) check_output(cmd, shell=True, stderr=subprocess.STDOUT) cmd = 'cd %s; find . -print0 | cpio --null -pdlu %s' % (src, dst) check_output(cmd, shell=True, stderr=subprocess.STDOUT) else: copytree(src, dst) def remove(path, recurse=True): """Equivalent to rm -f or rm -rf""" for name in glob.glob(path): try: os.unlink(name) except OSError as exc: if recurse and exc.errno == errno.EISDIR: shutil.rmtree(name) elif exc.errno != errno.ENOENT: raise def symlink(source, destination, force=False): """Create a symbolic link""" try: if force: remove(destination) os.symlink(source, destination) except OSError as e: if e.errno != errno.EEXIST or os.readlink(destination) != source: raise class CalledProcessError(Exception): def __init__(self, retcode, cmd, output = None): self.retcode = retcode self.cmd = cmd self.output = output def __str__(self): return "Command '%s' returned non-zero exit status %d with output %s" % (self.cmd, self.retcode, self.output) # Not needed when we move to python 2.7 def check_output(*popenargs, **kwargs): r"""Run command with arguments and return its output as a byte string. If the exit code was non-zero it raises a CalledProcessError. The CalledProcessError object will have the return code in the returncode attribute and output in the output attribute. The arguments are the same as for the Popen constructor. Example: >>> check_output(["ls", "-l", "/dev/null"]) 'crw-rw-rw- 1 root root 1, 3 Oct 18 2007 /dev/null\n' The stdout argument is not allowed as it is used internally. To capture standard error in the result, use stderr=STDOUT. >>> check_output(["/bin/sh", "-c", ... "ls -l non_existent_file ; exit 0"], ... stderr=STDOUT) 'ls: non_existent_file: No such file or directory\n' """ if 'stdout' in kwargs: raise ValueError('stdout argument not allowed, it will be overridden.') process = subprocess.Popen(stdout=subprocess.PIPE, *popenargs, **kwargs) output, unused_err = process.communicate() retcode = process.poll() if retcode: cmd = kwargs.get("args") if cmd is None: cmd = popenargs[0] raise CalledProcessError(retcode, cmd, output=output) return output def find(dir, **walkoptions): """ Given a directory, recurses into that directory, returning all files as absolute paths. """ for root, dirs, files in os.walk(dir, **walkoptions): for file in files: yield os.path.join(root, file) ## realpath() related functions def __is_path_below(file, root): return (file + os.path.sep).startswith(root) def __realpath_rel(start, rel_path, root, loop_cnt, assume_dir): """Calculates real path of symlink 'start' + 'rel_path' below 'root'; no part of 'start' below 'root' must contain symlinks. """ have_dir = True for d in rel_path.split(os.path.sep): if not have_dir and not assume_dir: raise OSError(errno.ENOENT, "no such directory %s" % start) if d == os.path.pardir: # '..' if len(start) >= len(root): # do not follow '..' before root start = os.path.dirname(start) else: # emit warning? pass else: (start, have_dir) = __realpath(os.path.join(start, d), root, loop_cnt, assume_dir) assert(__is_path_below(start, root)) return start def __realpath(file, root, loop_cnt, assume_dir): while os.path.islink(file) and len(file) >= len(root): if loop_cnt == 0: raise OSError(errno.ELOOP, file) loop_cnt -= 1 target = os.path.normpath(os.readlink(file)) if not os.path.isabs(target): tdir = os.path.dirname(file) assert(__is_path_below(tdir, root)) else: tdir = root file = __realpath_rel(tdir, target, root, loop_cnt, assume_dir) try: is_dir = os.path.isdir(file) except: is_dir = false return (file, is_dir) def realpath(file, root, use_physdir = True, loop_cnt = 100, assume_dir = False): """ Returns the canonical path of 'file' with assuming a toplevel 'root' directory. When 'use_physdir' is set, all preceding path components of 'file' will be resolved first; this flag should be set unless it is guaranteed that there is no symlink in the path. When 'assume_dir' is not set, missing path components will raise an ENOENT error""" root = os.path.normpath(root) file = os.path.normpath(file) if not root.endswith(os.path.sep): # letting root end with '/' makes some things easier root = root + os.path.sep if not __is_path_below(file, root): raise OSError(errno.EINVAL, "file '%s' is not below root" % file) try: if use_physdir: file = __realpath_rel(root, file[(len(root) - 1):], root, loop_cnt, assume_dir) else: file = __realpath(file, root, loop_cnt, assume_dir)[0] except OSError as e: if e.errno == errno.ELOOP: # make ELOOP more readable; without catching it, there will # be printed a backtrace with 100s of OSError exceptions # else raise OSError(errno.ELOOP, "too much recursions while resolving '%s'; loop in '%s'" % (file, e.strerror)) raise return file ########NEW FILE######## __FILENAME__ = prservice def prserv_make_conn(d, check = False): import prserv.serv host_params = filter(None, (d.getVar("PRSERV_HOST", True) or '').split(':')) try: conn = None conn = prserv.serv.PRServerConnection(host_params[0], int(host_params[1])) if check: if not conn.ping(): raise Exception('service not available') d.setVar("__PRSERV_CONN",conn) except Exception, exc: bb.fatal("Connecting to PR service %s:%s failed: %s" % (host_params[0], host_params[1], str(exc))) return conn def prserv_dump_db(d): if not d.getVar('PRSERV_HOST', True): bb.error("Not using network based PR service") return None conn = d.getVar("__PRSERV_CONN", True) if conn is None: conn = prserv_make_conn(d) if conn is None: bb.error("Making connection failed to remote PR service") return None #dump db opt_version = d.getVar('PRSERV_DUMPOPT_VERSION', True) opt_pkgarch = d.getVar('PRSERV_DUMPOPT_PKGARCH', True) opt_checksum = d.getVar('PRSERV_DUMPOPT_CHECKSUM', True) opt_col = ("1" == d.getVar('PRSERV_DUMPOPT_COL', True)) return conn.export(opt_version, opt_pkgarch, opt_checksum, opt_col) def prserv_import_db(d, filter_version=None, filter_pkgarch=None, filter_checksum=None): if not d.getVar('PRSERV_HOST', True): bb.error("Not using network based PR service") return None conn = d.getVar("__PRSERV_CONN", True) if conn is None: conn = prserv_make_conn(d) if conn is None: bb.error("Making connection failed to remote PR service") return None #get the entry values imported = [] prefix = "PRAUTO$" for v in d.keys(): if v.startswith(prefix): (remain, sep, checksum) = v.rpartition('$') (remain, sep, pkgarch) = remain.rpartition('$') (remain, sep, version) = remain.rpartition('$') if (remain + '$' != prefix) or \ (filter_version and filter_version != version) or \ (filter_pkgarch and filter_pkgarch != pkgarch) or \ (filter_checksum and filter_checksum != checksum): continue try: value = int(d.getVar(remain + '$' + version + '$' + pkgarch + '$' + checksum, True)) except BaseException as exc: bb.debug("Not valid value of %s:%s" % (v,str(exc))) continue ret = conn.importone(version,pkgarch,checksum,value) if ret != value: bb.error("importing(%s,%s,%s,%d) failed. DB may have larger value %d" % (version,pkgarch,checksum,value,ret)) else: imported.append((version,pkgarch,checksum,value)) return imported def prserv_export_tofile(d, metainfo, datainfo, lockdown, nomax=False): import bb.utils #initilize the output file bb.utils.mkdirhier(d.getVar('PRSERV_DUMPDIR', True)) df = d.getVar('PRSERV_DUMPFILE', True) #write data lf = bb.utils.lockfile("%s.lock" % df) f = open(df, "a") if metainfo: #dump column info f.write("#PR_core_ver = \"%s\"\n\n" % metainfo['core_ver']); f.write("#Table: %s\n" % metainfo['tbl_name']) f.write("#Columns:\n") f.write("#name \t type \t notn \t dflt \t pk\n") f.write("#----------\t --------\t --------\t --------\t ----\n") for i in range(len(metainfo['col_info'])): f.write("#%10s\t %8s\t %8s\t %8s\t %4s\n" % (metainfo['col_info'][i]['name'], metainfo['col_info'][i]['type'], metainfo['col_info'][i]['notnull'], metainfo['col_info'][i]['dflt_value'], metainfo['col_info'][i]['pk'])) f.write("\n") if lockdown: f.write("PRSERV_LOCKDOWN = \"1\"\n\n") if datainfo: idx = {} for i in range(len(datainfo)): pkgarch = datainfo[i]['pkgarch'] value = datainfo[i]['value'] if pkgarch not in idx: idx[pkgarch] = i elif value > datainfo[idx[pkgarch]]['value']: idx[pkgarch] = i f.write("PRAUTO$%s$%s$%s = \"%s\"\n" % (str(datainfo[i]['version']), pkgarch, str(datainfo[i]['checksum']), str(value))) if not nomax: for i in idx: f.write("PRAUTO_%s_%s = \"%s\"\n" % (str(datainfo[idx[i]]['version']),str(datainfo[idx[i]]['pkgarch']),str(datainfo[idx[i]]['value']))) f.close() bb.utils.unlockfile(lf) def prserv_check_avail(d): host_params = filter(None, (d.getVar("PRSERV_HOST", True) or '').split(':')) try: if len(host_params) != 2: raise TypeError else: int(host_params[1]) except TypeError: bb.fatal('Undefined/incorrect PRSERV_HOST value. Format: "host:port"') else: prserv_make_conn(d, True) ########NEW FILE######## __FILENAME__ = qa class ELFFile: EI_NIDENT = 16 EI_CLASS = 4 EI_DATA = 5 EI_VERSION = 6 EI_OSABI = 7 EI_ABIVERSION = 8 # possible values for EI_CLASS ELFCLASSNONE = 0 ELFCLASS32 = 1 ELFCLASS64 = 2 # possible value for EI_VERSION EV_CURRENT = 1 # possible values for EI_DATA ELFDATANONE = 0 ELFDATA2LSB = 1 ELFDATA2MSB = 2 def my_assert(self, expectation, result): if not expectation == result: #print "'%x','%x' %s" % (ord(expectation), ord(result), self.name) raise Exception("This does not work as expected") def __init__(self, name, bits = 0): self.name = name self.bits = bits self.objdump_output = {} def open(self): self.file = file(self.name, "r") self.data = self.file.read(ELFFile.EI_NIDENT+4) self.my_assert(len(self.data), ELFFile.EI_NIDENT+4) self.my_assert(self.data[0], chr(0x7f) ) self.my_assert(self.data[1], 'E') self.my_assert(self.data[2], 'L') self.my_assert(self.data[3], 'F') if self.bits == 0: if self.data[ELFFile.EI_CLASS] == chr(ELFFile.ELFCLASS32): self.bits = 32 elif self.data[ELFFile.EI_CLASS] == chr(ELFFile.ELFCLASS64): self.bits = 64 else: # Not 32-bit or 64.. lets assert raise Exception("ELF but not 32 or 64 bit.") elif self.bits == 32: self.my_assert(self.data[ELFFile.EI_CLASS], chr(ELFFile.ELFCLASS32)) elif self.bits == 64: self.my_assert(self.data[ELFFile.EI_CLASS], chr(ELFFile.ELFCLASS64)) else: raise Exception("Must specify unknown, 32 or 64 bit size.") self.my_assert(self.data[ELFFile.EI_VERSION], chr(ELFFile.EV_CURRENT) ) self.sex = self.data[ELFFile.EI_DATA] if self.sex == chr(ELFFile.ELFDATANONE): raise Exception("self.sex == ELFDATANONE") elif self.sex == chr(ELFFile.ELFDATA2LSB): self.sex = "<" elif self.sex == chr(ELFFile.ELFDATA2MSB): self.sex = ">" else: raise Exception("Unknown self.sex") def osAbi(self): return ord(self.data[ELFFile.EI_OSABI]) def abiVersion(self): return ord(self.data[ELFFile.EI_ABIVERSION]) def abiSize(self): return self.bits def isLittleEndian(self): return self.sex == "<" def isBigEngian(self): return self.sex == ">" def machine(self): """ We know the sex stored in self.sex and we know the position """ import struct (a,) = struct.unpack(self.sex+"H", self.data[18:20]) return a def run_objdump(self, cmd, d): import bb.process import sys if cmd in self.objdump_output: return self.objdump_output[cmd] objdump = d.getVar('OBJDUMP', True) env = os.environ.copy() env["LC_ALL"] = "C" env["PATH"] = d.getVar('PATH', True) try: bb.note("%s %s %s" % (objdump, cmd, self.name)) self.objdump_output[cmd] = bb.process.run([objdump, cmd, self.name], env=env, shell=False)[0] return self.objdump_output[cmd] except Exception as e: bb.note("%s %s %s failed: %s" % (objdump, cmd, self.name, e)) return "" ########NEW FILE######## __FILENAME__ = rootfs from abc import ABCMeta, abstractmethod from oe.utils import execute_pre_post_process from oe.package_manager import * from oe.manifest import * import oe.path import filecmp import shutil import os import subprocess import re class Rootfs(object): """ This is an abstract class. Do not instantiate this directly. """ __metaclass__ = ABCMeta def __init__(self, d): self.d = d self.pm = None self.image_rootfs = self.d.getVar('IMAGE_ROOTFS', True) self.deploy_dir_image = self.d.getVar('DEPLOY_DIR_IMAGE', True) self.install_order = Manifest.INSTALL_ORDER @abstractmethod def _create(self): pass @abstractmethod def _get_delayed_postinsts(self): pass @abstractmethod def _save_postinsts(self): pass @abstractmethod def _log_check(self): pass def _insert_feed_uris(self): if bb.utils.contains("IMAGE_FEATURES", "package-management", True, False, self.d): self.pm.insert_feeds_uris() @abstractmethod def _handle_intercept_failure(self, failed_script): pass """ The _cleanup() method should be used to clean-up stuff that we don't really want to end up on target. For example, in the case of RPM, the DB locks. The method is called, once, at the end of create() method. """ @abstractmethod def _cleanup(self): pass def _exec_shell_cmd(self, cmd): fakerootcmd = self.d.getVar('FAKEROOT', True) if fakerootcmd is not None: exec_cmd = [fakerootcmd, cmd] else: exec_cmd = cmd try: subprocess.check_output(exec_cmd, stderr=subprocess.STDOUT) except subprocess.CalledProcessError as e: return("Command '%s' returned %d:\n%s" % (e.cmd, e.returncode, e.output)) return None def create(self): bb.note("###### Generate rootfs #######") pre_process_cmds = self.d.getVar("ROOTFS_PREPROCESS_COMMAND", True) post_process_cmds = self.d.getVar("ROOTFS_POSTPROCESS_COMMAND", True) intercepts_dir = os.path.join(self.d.getVar('WORKDIR', True), "intercept_scripts") bb.utils.remove(intercepts_dir, True) bb.utils.mkdirhier(self.image_rootfs) bb.utils.mkdirhier(self.deploy_dir_image) shutil.copytree(self.d.expand("${COREBASE}/scripts/postinst-intercepts"), intercepts_dir) shutil.copy(self.d.expand("${COREBASE}/meta/files/deploydir_readme.txt"), self.deploy_dir_image + "/README_-_DO_NOT_DELETE_FILES_IN_THIS_DIRECTORY.txt") execute_pre_post_process(self.d, pre_process_cmds) # call the package manager dependent create method self._create() sysconfdir = self.image_rootfs + self.d.getVar('sysconfdir', True) bb.utils.mkdirhier(sysconfdir) with open(sysconfdir + "/version", "w+") as ver: ver.write(self.d.getVar('BUILDNAME', True) + "\n") self._run_intercepts() execute_pre_post_process(self.d, post_process_cmds) if bb.utils.contains("IMAGE_FEATURES", "read-only-rootfs", True, False, self.d): delayed_postinsts = self._get_delayed_postinsts() if delayed_postinsts is not None: bb.fatal("The following packages could not be configured" "offline and rootfs is read-only: %s" % delayed_postinsts) if self.d.getVar('USE_DEVFS', True) != "1": self._create_devfs() self._uninstall_uneeded() self._insert_feed_uris() self._run_ldconfig() self._generate_kernel_module_deps() self._cleanup() def _uninstall_uneeded(self): if bb.utils.contains("IMAGE_FEATURES", "package-management", True, False, self.d): return delayed_postinsts = self._get_delayed_postinsts() if delayed_postinsts is None: installed_pkgs_dir = self.d.expand('${WORKDIR}/installed_pkgs.txt') pkgs_to_remove = list() with open(installed_pkgs_dir, "r+") as installed_pkgs: pkgs_installed = installed_pkgs.read().split('\n') for pkg_installed in pkgs_installed[:]: pkg = pkg_installed.split()[0] if pkg in ["update-rc.d", "base-passwd", self.d.getVar("ROOTFS_BOOTSTRAP_INSTALL", True) ]: pkgs_to_remove.append(pkg) pkgs_installed.remove(pkg_installed) if len(pkgs_to_remove) > 0: self.pm.remove(pkgs_to_remove, False) # Update installed_pkgs.txt open(installed_pkgs_dir, "w+").write('\n'.join(pkgs_installed)) if os.path.exists(self.d.expand("${IMAGE_ROOTFS}${sysconfdir}/init.d/run-postinsts")): self._exec_shell_cmd(["update-rc.d", "-f", "-r", self.d.getVar('IMAGE_ROOTFS', True), "run-postinsts", "remove"]) else: self._save_postinsts() self.pm.remove_packaging_data() def _run_intercepts(self): intercepts_dir = os.path.join(self.d.getVar('WORKDIR', True), "intercept_scripts") bb.note("Running intercept scripts:") os.environ['D'] = self.image_rootfs for script in os.listdir(intercepts_dir): script_full = os.path.join(intercepts_dir, script) if script == "postinst_intercept" or not os.access(script_full, os.X_OK): continue bb.note("> Executing %s intercept ..." % script) try: subprocess.check_output(script_full) except subprocess.CalledProcessError as e: bb.warn("The postinstall intercept hook '%s' failed (exit code: %d)! See log for details!" % (script, e.returncode)) with open(script_full) as intercept: registered_pkgs = None for line in intercept.read().split("\n"): m = re.match("^##PKGS:(.*)", line) if m is not None: registered_pkgs = m.group(1).strip() break if registered_pkgs is not None: bb.warn("The postinstalls for the following packages " "will be postponed for first boot: %s" % registered_pkgs) # call the backend dependent handler self._handle_intercept_failure(registered_pkgs) def _run_ldconfig(self): if self.d.getVar('LDCONFIGDEPEND', True): bb.note("Executing: ldconfig -r" + self.image_rootfs + "-c new -v") self._exec_shell_cmd(['ldconfig', '-r', self.image_rootfs, '-c', 'new', '-v']) def _generate_kernel_module_deps(self): kernel_abi_ver_file = os.path.join(self.d.getVar('STAGING_KERNEL_DIR', True), 'kernel-abiversion') if os.path.exists(kernel_abi_ver_file): kernel_ver = open(kernel_abi_ver_file).read().strip(' \n') modules_dir = os.path.join(self.image_rootfs, 'lib', 'modules', kernel_ver) bb.utils.mkdirhier(modules_dir) self._exec_shell_cmd(['depmodwrapper', '-a', '-b', self.image_rootfs, kernel_ver]) """ Create devfs: * IMAGE_DEVICE_TABLE is the old name to an absolute path to a device table file * IMAGE_DEVICE_TABLES is a new name for a file, or list of files, seached for in the BBPATH If neither are specified then the default name of files/device_table-minimal.txt is searched for in the BBPATH (same as the old version.) """ def _create_devfs(self): devtable_list = [] devtable = self.d.getVar('IMAGE_DEVICE_TABLE', True) if devtable is not None: devtable_list.append(devtable) else: devtables = self.d.getVar('IMAGE_DEVICE_TABLES', True) if devtables is None: devtables = 'files/device_table-minimal.txt' for devtable in devtables.split(): devtable_list.append("%s" % bb.utils.which(self.d.getVar('BBPATH', True), devtable)) for devtable in devtable_list: self._exec_shell_cmd(["makedevs", "-r", self.image_rootfs, "-D", devtable]) class RpmRootfs(Rootfs): def __init__(self, d, manifest_dir): super(RpmRootfs, self).__init__(d) self.manifest = RpmManifest(d, manifest_dir) self.pm = RpmPM(d, d.getVar('IMAGE_ROOTFS', True), self.d.getVar('TARGET_VENDOR', True) ) self.inc_rpm_image_gen = self.d.getVar('INC_RPM_IMAGE_GEN', True) if self.inc_rpm_image_gen != "1": bb.utils.remove(self.image_rootfs, True) else: self.pm.recovery_packaging_data() bb.utils.remove(self.d.getVar('MULTILIB_TEMP_ROOTFS', True), True) self.pm.create_configs() ''' While rpm incremental image generation is enabled, it will remove the unneeded pkgs by comparing the new install solution manifest and the old installed manifest. ''' def _create_incremental(self, pkgs_initial_install): if self.inc_rpm_image_gen == "1": pkgs_to_install = list() for pkg_type in pkgs_initial_install: pkgs_to_install += pkgs_initial_install[pkg_type] installed_manifest = self.pm.load_old_install_solution() solution_manifest = self.pm.dump_install_solution(pkgs_to_install) pkg_to_remove = list() for pkg in installed_manifest: if pkg not in solution_manifest: pkg_to_remove.append(pkg) self.pm.update() bb.note('incremental update -- upgrade packages in place ') self.pm.upgrade() if pkg_to_remove != []: bb.note('incremental removed: %s' % ' '.join(pkg_to_remove)) self.pm.remove(pkg_to_remove) def _create(self): pkgs_to_install = self.manifest.parse_initial_manifest() # update PM index files self.pm.write_index() self.pm.dump_all_available_pkgs() if self.inc_rpm_image_gen == "1": self._create_incremental(pkgs_to_install) self.pm.update() pkgs = [] pkgs_attempt = [] for pkg_type in pkgs_to_install: if pkg_type == Manifest.PKG_TYPE_ATTEMPT_ONLY: pkgs_attempt += pkgs_to_install[pkg_type] else: pkgs += pkgs_to_install[pkg_type] self.pm.install(pkgs) self.pm.install(pkgs_attempt, True) self.pm.install_complementary() self._log_check() if self.inc_rpm_image_gen == "1": self.pm.backup_packaging_data() self.pm.rpm_setup_smart_target_config() def _get_delayed_postinsts(self): postinst_dir = self.d.expand("${IMAGE_ROOTFS}${sysconfdir}/rpm-postinsts") if os.path.isdir(postinst_dir): files = os.listdir(postinst_dir) for f in files: bb.note('Delayed package scriptlet: %s' % f) return files return None def _save_postinsts(self): # this is just a stub. For RPM, the failed postinstalls are # already saved in /etc/rpm-postinsts pass def _log_check(self): r = re.compile('(unpacking of archive failed|Cannot find package|exit 1|ERR|Fail)') log_path = self.d.expand("${T}/log.do_rootfs") with open(log_path, 'r') as log: found_error = 0 message = "\n" for line in log.read().split('\n'): if 'log_check' in line: continue m = r.search(line) if m: found_error = 1 bb.warn('log_check: There were error messages in the logfile') bb.warn('log_check: Matched keyword: [%s]\n\n' % m.group()) if found_error >= 1 and found_error <= 5: message += line + '\n' found_error += 1 if found_error == 6: bb.fatal(message) def _handle_intercept_failure(self, registered_pkgs): rpm_postinsts_dir = self.image_rootfs + self.d.expand('${sysconfdir}/rpm-postinsts/') bb.utils.mkdirhier(rpm_postinsts_dir) # Save the package postinstalls in /etc/rpm-postinsts for pkg in registered_pkgs.split(): self.pm.save_rpmpostinst(pkg) def _cleanup(self): # during the execution of postprocess commands, rpm is called several # times to get the files installed, dependencies, etc. This creates the # __db.00* (Berkeley DB files that hold locks, rpm specific environment # settings, etc.), that should not get into the final rootfs self.pm.unlock_rpm_db() class DpkgRootfs(Rootfs): def __init__(self, d, manifest_dir): super(DpkgRootfs, self).__init__(d) bb.utils.remove(self.image_rootfs, True) bb.utils.remove(self.d.getVar('MULTILIB_TEMP_ROOTFS', True), True) self.manifest = DpkgManifest(d, manifest_dir) self.pm = DpkgPM(d, d.getVar('IMAGE_ROOTFS', True), d.getVar('PACKAGE_ARCHS', True), d.getVar('DPKG_ARCH', True)) def _create(self): pkgs_to_install = self.manifest.parse_initial_manifest() alt_dir = self.d.expand("${IMAGE_ROOTFS}/var/lib/dpkg/alternatives") bb.utils.mkdirhier(alt_dir) # update PM index files self.pm.write_index() self.pm.update() for pkg_type in self.install_order: if pkg_type in pkgs_to_install: self.pm.install(pkgs_to_install[pkg_type], [False, True][pkg_type == Manifest.PKG_TYPE_ATTEMPT_ONLY]) self.pm.install_complementary() self.pm.fix_broken_dependencies() self.pm.mark_packages("installed") self.pm.run_pre_post_installs() def _get_delayed_postinsts(self): pkg_list = [] with open(self.image_rootfs + "/var/lib/dpkg/status") as status: for line in status: m_pkg = re.match("^Package: (.*)", line) m_status = re.match("^Status:.*unpacked", line) if m_pkg is not None: pkg_name = m_pkg.group(1) elif m_status is not None: pkg_list.append(pkg_name) if len(pkg_list) == 0: return None return pkg_list def _save_postinsts(self): num = 0 for p in self._get_delayed_postinsts(): dst_postinst_dir = self.d.expand("${IMAGE_ROOTFS}${sysconfdir}/deb-postinsts") src_postinst_dir = self.d.expand("${IMAGE_ROOTFS}/var/lib/dpkg/info") bb.utils.mkdirhier(dst_postinst_dir) if os.path.exists(os.path.join(src_postinst_dir, p + ".postinst")): shutil.copy(os.path.join(src_postinst_dir, p + ".postinst"), os.path.join(dst_postinst_dir, "%03d-%s" % (num, p))) num += 1 def _handle_intercept_failure(self, registered_pkgs): self.pm.mark_packages("unpacked", registered_pkgs.split()) def _log_check(self): pass def _cleanup(self): pass class OpkgRootfs(Rootfs): def __init__(self, d, manifest_dir): super(OpkgRootfs, self).__init__(d) self.manifest = OpkgManifest(d, manifest_dir) self.opkg_conf = self.d.getVar("IPKGCONF_TARGET", True) self.pkg_archs = self.d.getVar("ALL_MULTILIB_PACKAGE_ARCHS", True) self.inc_opkg_image_gen = self.d.getVar('INC_IPK_IMAGE_GEN', True) or "" if self._remove_old_rootfs(): bb.utils.remove(self.image_rootfs, True) self.pm = OpkgPM(d, self.image_rootfs, self.opkg_conf, self.pkg_archs) else: self.pm = OpkgPM(d, self.image_rootfs, self.opkg_conf, self.pkg_archs) self.pm.recover_packaging_data() bb.utils.remove(self.d.getVar('MULTILIB_TEMP_ROOTFS', True), True) def _prelink_file(self, root_dir, filename): bb.note('prelink %s in %s' % (filename, root_dir)) prelink_cfg = oe.path.join(root_dir, self.d.expand('${sysconfdir}/prelink.conf')) if not os.path.exists(prelink_cfg): shutil.copy(self.d.expand('${STAGING_DIR_NATIVE}${sysconfdir_native}/prelink.conf'), prelink_cfg) cmd_prelink = self.d.expand('${STAGING_DIR_NATIVE}${sbindir_native}/prelink') self._exec_shell_cmd([cmd_prelink, '--root', root_dir, '-amR', '-N', '-c', self.d.expand('${sysconfdir}/prelink.conf')]) ''' Compare two files with the same key twice to see if they are equal. If they are not equal, it means they are duplicated and come from different packages. 1st: Comapre them directly; 2nd: While incremental image creation is enabled, one of the files could be probaly prelinked in the previous image creation and the file has been changed, so we need to prelink the other one and compare them. ''' def _file_equal(self, key, f1, f2): # Both of them are not prelinked if filecmp.cmp(f1, f2): return True if self.image_rootfs not in f1: self._prelink_file(f1.replace(key, ''), f1) if self.image_rootfs not in f2: self._prelink_file(f2.replace(key, ''), f2) # Both of them are prelinked if filecmp.cmp(f1, f2): return True # Not equal return False """ This function was reused from the old implementation. See commit: "image.bbclass: Added variables for multilib support." by Lianhao Lu. """ def _multilib_sanity_test(self, dirs): allow_replace = self.d.getVar("MULTILIBRE_ALLOW_REP", True) if allow_replace is None: allow_replace = "" allow_rep = re.compile(re.sub("\|$", "", allow_replace)) error_prompt = "Multilib check error:" files = {} for dir in dirs: for root, subfolders, subfiles in os.walk(dir): for file in subfiles: item = os.path.join(root, file) key = str(os.path.join("/", os.path.relpath(item, dir))) valid = True if key in files: #check whether the file is allow to replace if allow_rep.match(key): valid = True else: if os.path.exists(files[key]) and \ os.path.exists(item) and \ not self._file_equal(key, files[key], item): valid = False bb.fatal("%s duplicate files %s %s is not the same\n" % (error_prompt, item, files[key])) #pass the check, add to list if valid: files[key] = item def _multilib_test_install(self, pkgs): ml_temp = self.d.getVar("MULTILIB_TEMP_ROOTFS", True) bb.utils.mkdirhier(ml_temp) dirs = [self.image_rootfs] for variant in self.d.getVar("MULTILIB_VARIANTS", True).split(): ml_target_rootfs = os.path.join(ml_temp, variant) bb.utils.remove(ml_target_rootfs, True) ml_opkg_conf = os.path.join(ml_temp, variant + "-" + os.path.basename(self.opkg_conf)) ml_pm = OpkgPM(self.d, ml_target_rootfs, ml_opkg_conf, self.pkg_archs) ml_pm.update() ml_pm.install(pkgs) dirs.append(ml_target_rootfs) self._multilib_sanity_test(dirs) ''' While ipk incremental image generation is enabled, it will remove the unneeded pkgs by comparing the old full manifest in previous existing image and the new full manifest in the current image. ''' def _remove_extra_packages(self, pkgs_initial_install): if self.inc_opkg_image_gen == "1": # Parse full manifest in previous existing image creation session old_full_manifest = self.manifest.parse_full_manifest() # Create full manifest for the current image session, the old one # will be replaced by the new one. self.manifest.create_full(self.pm) # Parse full manifest in current image creation session new_full_manifest = self.manifest.parse_full_manifest() pkg_to_remove = list() for pkg in old_full_manifest: if pkg not in new_full_manifest: pkg_to_remove.append(pkg) if pkg_to_remove != []: bb.note('decremental removed: %s' % ' '.join(pkg_to_remove)) self.pm.remove(pkg_to_remove) ''' Compare with previous existing image creation, if some conditions triggered, the previous old image should be removed. The conditions include any of 'PACKAGE_EXCLUDE, NO_RECOMMENDATIONS and BAD_RECOMMENDATIONS' has been changed. ''' def _remove_old_rootfs(self): if self.inc_opkg_image_gen != "1": return True vars_list_file = self.d.expand('${T}/vars_list') old_vars_list = "" if os.path.exists(vars_list_file): old_vars_list = open(vars_list_file, 'r+').read() new_vars_list = '%s:%s:%s\n' % \ ((self.d.getVar('BAD_RECOMMENDATIONS', True) or '').strip(), (self.d.getVar('NO_RECOMMENDATIONS', True) or '').strip(), (self.d.getVar('PACKAGE_EXCLUDE', True) or '').strip()) open(vars_list_file, 'w+').write(new_vars_list) if old_vars_list != new_vars_list: return True return False def _create(self): pkgs_to_install = self.manifest.parse_initial_manifest() opkg_pre_process_cmds = self.d.getVar('OPKG_PREPROCESS_COMMANDS', True) opkg_post_process_cmds = self.d.getVar('OPKG_POSTPROCESS_COMMANDS', True) rootfs_post_install_cmds = self.d.getVar('ROOTFS_POSTINSTALL_COMMAND', True) # update PM index files, unless users provide their own feeds if (self.d.getVar('BUILD_IMAGES_FROM_FEEDS', True) or "") != "1": self.pm.write_index() execute_pre_post_process(self.d, opkg_pre_process_cmds) self.pm.update() self.pm.handle_bad_recommendations() if self.inc_opkg_image_gen == "1": self._remove_extra_packages(pkgs_to_install) for pkg_type in self.install_order: if pkg_type in pkgs_to_install: # For multilib, we perform a sanity test before final install # If sanity test fails, it will automatically do a bb.fatal() # and the installation will stop if pkg_type == Manifest.PKG_TYPE_MULTILIB: self._multilib_test_install(pkgs_to_install[pkg_type]) self.pm.install(pkgs_to_install[pkg_type], [False, True][pkg_type == Manifest.PKG_TYPE_ATTEMPT_ONLY]) self.pm.install_complementary() execute_pre_post_process(self.d, opkg_post_process_cmds) execute_pre_post_process(self.d, rootfs_post_install_cmds) if self.inc_opkg_image_gen == "1": self.pm.backup_packaging_data() def _get_delayed_postinsts(self): pkg_list = [] status_file = os.path.join(self.image_rootfs, self.d.getVar('OPKGLIBDIR', True).strip('/'), "opkg", "status") with open(status_file) as status: for line in status: m_pkg = re.match("^Package: (.*)", line) m_status = re.match("^Status:.*unpacked", line) if m_pkg is not None: pkg_name = m_pkg.group(1) elif m_status is not None: pkg_list.append(pkg_name) if len(pkg_list) == 0: return None return pkg_list def _save_postinsts(self): num = 0 for p in self._get_delayed_postinsts(): dst_postinst_dir = self.d.expand("${IMAGE_ROOTFS}${sysconfdir}/ipk-postinsts") src_postinst_dir = self.d.expand("${IMAGE_ROOTFS}${OPKGLIBDIR}/opkg/info") bb.utils.mkdirhier(dst_postinst_dir) if os.path.exists(os.path.join(src_postinst_dir, p + ".postinst")): shutil.copy(os.path.join(src_postinst_dir, p + ".postinst"), os.path.join(dst_postinst_dir, "%03d-%s" % (num, p))) num += 1 def _handle_intercept_failure(self, registered_pkgs): self.pm.mark_packages("unpacked", registered_pkgs.split()) def _log_check(self): pass def _cleanup(self): pass def create_rootfs(d, manifest_dir=None): env_bkp = os.environ.copy() img_type = d.getVar('IMAGE_PKGTYPE', True) if img_type == "rpm": RpmRootfs(d, manifest_dir).create() elif img_type == "ipk": OpkgRootfs(d, manifest_dir).create() elif img_type == "deb": DpkgRootfs(d, manifest_dir).create() os.environ.clear() os.environ.update(env_bkp) def image_list_installed_packages(d, format=None, rootfs_dir=None): if not rootfs_dir: rootfs_dir = d.getVar('IMAGE_ROOTFS', True) img_type = d.getVar('IMAGE_PKGTYPE', True) if img_type == "rpm": return RpmPkgsList(d, rootfs_dir).list(format) elif img_type == "ipk": return OpkgPkgsList(d, rootfs_dir, d.getVar("IPKGCONF_TARGET", True)).list(format) elif img_type == "deb": return DpkgPkgsList(d, rootfs_dir).list(format) if __name__ == "__main__": """ We should be able to run this as a standalone script, from outside bitbake environment. """ """ TBD """ ########NEW FILE######## __FILENAME__ = sdk from abc import ABCMeta, abstractmethod from oe.utils import execute_pre_post_process from oe.manifest import * from oe.package_manager import * import os import shutil import glob class Sdk(object): __metaclass__ = ABCMeta def __init__(self, d, manifest_dir): self.d = d self.sdk_output = self.d.getVar('SDK_OUTPUT', True) self.sdk_native_path = self.d.getVar('SDKPATHNATIVE', True).strip('/') self.target_path = self.d.getVar('SDKTARGETSYSROOT', True).strip('/') self.sysconfdir = self.d.getVar('sysconfdir', True).strip('/') self.sdk_target_sysroot = os.path.join(self.sdk_output, self.target_path) self.sdk_host_sysroot = self.sdk_output if manifest_dir is None: self.manifest_dir = self.d.getVar("SDK_DIR", True) else: self.manifest_dir = manifest_dir bb.utils.remove(self.sdk_output, True) self.install_order = Manifest.INSTALL_ORDER @abstractmethod def _populate(self): pass def populate(self): bb.utils.mkdirhier(self.sdk_output) # call backend dependent implementation self._populate() # Don't ship any libGL in the SDK bb.utils.remove(os.path.join(self.sdk_output, self.sdk_native_path, self.d.getVar('libdir_nativesdk', True).strip('/'), "libGL*")) # Fix or remove broken .la files bb.utils.remove(os.path.join(self.sdk_output, self.sdk_native_path, self.d.getVar('libdir_nativesdk', True).strip('/'), "*.la")) # Link the ld.so.cache file into the hosts filesystem link_name = os.path.join(self.sdk_output, self.sdk_native_path, self.sysconfdir, "ld.so.cache") bb.utils.mkdirhier(os.path.dirname(link_name)) os.symlink("/etc/ld.so.cache", link_name) execute_pre_post_process(self.d, self.d.getVar('SDK_POSTPROCESS_COMMAND', True)) class RpmSdk(Sdk): def __init__(self, d, manifest_dir=None): super(RpmSdk, self).__init__(d, manifest_dir) self.target_manifest = RpmManifest(d, self.manifest_dir, Manifest.MANIFEST_TYPE_SDK_TARGET) self.host_manifest = RpmManifest(d, self.manifest_dir, Manifest.MANIFEST_TYPE_SDK_HOST) target_providename = ['/bin/sh', '/bin/bash', '/usr/bin/env', '/usr/bin/perl', 'pkgconfig' ] self.target_pm = RpmPM(d, self.sdk_target_sysroot, self.d.getVar('TARGET_VENDOR', True), 'target', target_providename ) sdk_providename = ['/bin/sh', '/bin/bash', '/usr/bin/env', '/usr/bin/perl', 'pkgconfig', 'libGL.so()(64bit)', 'libGL.so' ] self.host_pm = RpmPM(d, self.sdk_host_sysroot, self.d.getVar('SDK_VENDOR', True), 'host', sdk_providename, "SDK_PACKAGE_ARCHS", "SDK_OS" ) def _populate_sysroot(self, pm, manifest): pkgs_to_install = manifest.parse_initial_manifest() pm.create_configs() pm.write_index() pm.dump_all_available_pkgs() pm.update() for pkg_type in self.install_order: if pkg_type in pkgs_to_install: pm.install(pkgs_to_install[pkg_type], [False, True][pkg_type == Manifest.PKG_TYPE_ATTEMPT_ONLY]) def _populate(self): bb.note("Installing TARGET packages") self._populate_sysroot(self.target_pm, self.target_manifest) self.target_pm.install_complementary(self.d.getVar('SDKIMAGE_INSTALL_COMPLEMENTARY', True)) execute_pre_post_process(self.d, self.d.getVar("POPULATE_SDK_POST_TARGET_COMMAND", True)) self.target_pm.remove_packaging_data() bb.note("Installing NATIVESDK packages") self._populate_sysroot(self.host_pm, self.host_manifest) execute_pre_post_process(self.d, self.d.getVar("POPULATE_SDK_POST_HOST_COMMAND", True)) self.host_pm.remove_packaging_data() # Move host RPM library data native_rpm_state_dir = os.path.join(self.sdk_output, self.sdk_native_path, self.d.getVar('localstatedir_nativesdk', True).strip('/'), "lib", "rpm" ) bb.utils.mkdirhier(native_rpm_state_dir) for f in glob.glob(os.path.join(self.sdk_output, "var", "lib", "rpm", "*")): bb.utils.movefile(f, native_rpm_state_dir) bb.utils.remove(os.path.join(self.sdk_output, "var"), True) # Move host sysconfig data native_sysconf_dir = os.path.join(self.sdk_output, self.sdk_native_path, self.d.getVar('sysconfdir', True).strip('/'), ) bb.utils.mkdirhier(native_sysconf_dir) for f in glob.glob(os.path.join(self.sdk_output, "etc", "*")): bb.utils.movefile(f, native_sysconf_dir) bb.utils.remove(os.path.join(self.sdk_output, "etc"), True) class OpkgSdk(Sdk): def __init__(self, d, manifest_dir=None): super(OpkgSdk, self).__init__(d, manifest_dir) self.target_conf = self.d.getVar("IPKGCONF_TARGET", True) self.host_conf = self.d.getVar("IPKGCONF_SDK", True) self.target_manifest = OpkgManifest(d, self.manifest_dir, Manifest.MANIFEST_TYPE_SDK_TARGET) self.host_manifest = OpkgManifest(d, self.manifest_dir, Manifest.MANIFEST_TYPE_SDK_HOST) self.target_pm = OpkgPM(d, self.sdk_target_sysroot, self.target_conf, self.d.getVar("ALL_MULTILIB_PACKAGE_ARCHS", True)) self.host_pm = OpkgPM(d, self.sdk_host_sysroot, self.host_conf, self.d.getVar("SDK_PACKAGE_ARCHS", True)) def _populate_sysroot(self, pm, manifest): pkgs_to_install = manifest.parse_initial_manifest() if (self.d.getVar('BUILD_IMAGES_FROM_FEEDS', True) or "") != "1": pm.write_index() pm.update() for pkg_type in self.install_order: if pkg_type in pkgs_to_install: pm.install(pkgs_to_install[pkg_type], [False, True][pkg_type == Manifest.PKG_TYPE_ATTEMPT_ONLY]) def _populate(self): bb.note("Installing TARGET packages") self._populate_sysroot(self.target_pm, self.target_manifest) self.target_pm.install_complementary(self.d.getVar('SDKIMAGE_INSTALL_COMPLEMENTARY', True)) execute_pre_post_process(self.d, self.d.getVar("POPULATE_SDK_POST_TARGET_COMMAND", True)) bb.note("Installing NATIVESDK packages") self._populate_sysroot(self.host_pm, self.host_manifest) execute_pre_post_process(self.d, self.d.getVar("POPULATE_SDK_POST_HOST_COMMAND", True)) target_sysconfdir = os.path.join(self.sdk_target_sysroot, self.sysconfdir) host_sysconfdir = os.path.join(self.sdk_host_sysroot, self.sysconfdir) bb.utils.mkdirhier(target_sysconfdir) shutil.copy(self.target_conf, target_sysconfdir) os.chmod(os.path.join(target_sysconfdir, os.path.basename(self.target_conf)), 0644) bb.utils.mkdirhier(host_sysconfdir) shutil.copy(self.host_conf, host_sysconfdir) os.chmod(os.path.join(host_sysconfdir, os.path.basename(self.host_conf)), 0644) native_opkg_state_dir = os.path.join(self.sdk_output, self.sdk_native_path, self.d.getVar('localstatedir_nativesdk', True).strip('/'), "lib", "opkg") bb.utils.mkdirhier(native_opkg_state_dir) for f in glob.glob(os.path.join(self.sdk_output, "var", "lib", "opkg", "*")): bb.utils.movefile(f, native_opkg_state_dir) bb.utils.remove(os.path.join(self.sdk_output, "var"), True) class DpkgSdk(Sdk): def __init__(self, d, manifest_dir=None): super(DpkgSdk, self).__init__(d, manifest_dir) self.target_conf_dir = os.path.join(self.d.getVar("APTCONF_TARGET", True), "apt") self.host_conf_dir = os.path.join(self.d.getVar("APTCONF_TARGET", True), "apt-sdk") self.target_manifest = DpkgManifest(d, self.manifest_dir, Manifest.MANIFEST_TYPE_SDK_TARGET) self.host_manifest = DpkgManifest(d, self.manifest_dir, Manifest.MANIFEST_TYPE_SDK_HOST) self.target_pm = DpkgPM(d, self.sdk_target_sysroot, self.d.getVar("PACKAGE_ARCHS", True), self.d.getVar("DPKG_ARCH", True), self.target_conf_dir) self.host_pm = DpkgPM(d, self.sdk_host_sysroot, self.d.getVar("SDK_PACKAGE_ARCHS", True), self.d.getVar("DEB_SDK_ARCH", True), self.host_conf_dir) def _copy_apt_dir_to(self, dst_dir): staging_etcdir_native = self.d.getVar("STAGING_ETCDIR_NATIVE", True) bb.utils.remove(dst_dir, True) shutil.copytree(os.path.join(staging_etcdir_native, "apt"), dst_dir) def _populate_sysroot(self, pm, manifest): pkgs_to_install = manifest.parse_initial_manifest() pm.write_index() pm.update() for pkg_type in self.install_order: if pkg_type in pkgs_to_install: pm.install(pkgs_to_install[pkg_type], [False, True][pkg_type == Manifest.PKG_TYPE_ATTEMPT_ONLY]) def _populate(self): bb.note("Installing TARGET packages") self._populate_sysroot(self.target_pm, self.target_manifest) execute_pre_post_process(self.d, self.d.getVar("POPULATE_SDK_POST_TARGET_COMMAND", True)) self._copy_apt_dir_to(os.path.join(self.sdk_target_sysroot, "etc", "apt")) bb.note("Installing NATIVESDK packages") self._populate_sysroot(self.host_pm, self.host_manifest) execute_pre_post_process(self.d, self.d.getVar("POPULATE_SDK_POST_HOST_COMMAND", True)) self._copy_apt_dir_to(os.path.join(self.sdk_output, self.sdk_native_path, "etc", "apt")) native_dpkg_state_dir = os.path.join(self.sdk_output, self.sdk_native_path, "var", "lib", "dpkg") bb.utils.mkdirhier(native_dpkg_state_dir) for f in glob.glob(os.path.join(self.sdk_output, "var", "lib", "dpkg", "*")): bb.utils.movefile(f, native_dpkg_state_dir) bb.utils.remove(os.path.join(self.sdk_output, "var"), True) def sdk_list_installed_packages(d, target, format=None, rootfs_dir=None): if rootfs_dir is None: sdk_output = d.getVar('SDK_OUTPUT', True) target_path = d.getVar('SDKTARGETSYSROOT', True).strip('/') rootfs_dir = [sdk_output, os.path.join(sdk_output, target_path)][target is True] img_type = d.getVar('IMAGE_PKGTYPE', True) if img_type == "rpm": arch_var = ["SDK_PACKAGE_ARCHS", None][target is True] os_var = ["SDK_OS", None][target is True] return RpmPkgsList(d, rootfs_dir, arch_var, os_var).list(format) elif img_type == "ipk": conf_file_var = ["IPKGCONF_SDK", "IPKGCONF_Target"][target is True] return OpkgPkgsList(d, rootfs_dir, d.getVar(conf_file_var, True)).list(format) elif img_type == "deb": return DpkgPkgsList(d, rootfs_dir).list(format) def populate_sdk(d, manifest_dir=None): env_bkp = os.environ.copy() img_type = d.getVar('IMAGE_PKGTYPE', True) if img_type == "rpm": RpmSdk(d, manifest_dir).populate() elif img_type == "ipk": OpkgSdk(d, manifest_dir).populate() elif img_type == "deb": DpkgSdk(d, manifest_dir).populate() os.environ.clear() os.environ.update(env_bkp) if __name__ == "__main__": pass ########NEW FILE######## __FILENAME__ = sstatesig import bb.siggen def sstate_rundepfilter(siggen, fn, recipename, task, dep, depname, dataCache): # Return True if we should keep the dependency, False to drop it def isNative(x): return x.endswith("-native") def isCross(x): return "-cross-" in x def isNativeSDK(x): return x.startswith("nativesdk-") def isKernel(fn): inherits = " ".join(dataCache.inherits[fn]) return inherits.find("/module-base.bbclass") != -1 or inherits.find("/linux-kernel-base.bbclass") != -1 def isPackageGroup(fn): inherits = " ".join(dataCache.inherits[fn]) return "/packagegroup.bbclass" in inherits def isImage(fn): return "/image.bbclass" in " ".join(dataCache.inherits[fn]) # Always include our own inter-task dependencies if recipename == depname: return True # Quilt (patch application) changing isn't likely to affect anything excludelist = ['quilt-native', 'subversion-native', 'git-native'] if depname in excludelist and recipename != depname: return False # Exclude well defined recipe->dependency if "%s->%s" % (recipename, depname) in siggen.saferecipedeps: return False # Don't change native/cross/nativesdk recipe dependencies any further if isNative(recipename) or isCross(recipename) or isNativeSDK(recipename): return True # Only target packages beyond here # packagegroups are assumed to have well behaved names which don't change between architecures/tunes if isPackageGroup(fn): return False # Exclude well defined machine specific configurations which don't change ABI if depname in siggen.abisaferecipes and not isImage(fn): return False # Kernel modules are well namespaced. We don't want to depend on the kernel's checksum # if we're just doing an RRECOMMENDS_xxx = "kernel-module-*", not least because the checksum # is machine specific. # Therefore if we're not a kernel or a module recipe (inheriting the kernel classes) # and we reccomend a kernel-module, we exclude the dependency. depfn = dep.rsplit(".", 1)[0] if dataCache and isKernel(depfn) and not isKernel(fn): for pkg in dataCache.runrecs[fn]: if " ".join(dataCache.runrecs[fn][pkg]).find("kernel-module-") != -1: return False # Default to keep dependencies return True class SignatureGeneratorOEBasic(bb.siggen.SignatureGeneratorBasic): name = "OEBasic" def init_rundepcheck(self, data): self.abisaferecipes = (data.getVar("SIGGEN_EXCLUDERECIPES_ABISAFE", True) or "").split() self.saferecipedeps = (data.getVar("SIGGEN_EXCLUDE_SAFE_RECIPE_DEPS", True) or "").split() pass def rundep_check(self, fn, recipename, task, dep, depname, dataCache = None): return sstate_rundepfilter(self, fn, recipename, task, dep, depname, dataCache) class SignatureGeneratorOEBasicHash(bb.siggen.SignatureGeneratorBasicHash): name = "OEBasicHash" def init_rundepcheck(self, data): self.abisaferecipes = (data.getVar("SIGGEN_EXCLUDERECIPES_ABISAFE", True) or "").split() self.saferecipedeps = (data.getVar("SIGGEN_EXCLUDE_SAFE_RECIPE_DEPS", True) or "").split() pass def rundep_check(self, fn, recipename, task, dep, depname, dataCache = None): return sstate_rundepfilter(self, fn, recipename, task, dep, depname, dataCache) # Insert these classes into siggen's namespace so it can see and select them bb.siggen.SignatureGeneratorOEBasic = SignatureGeneratorOEBasic bb.siggen.SignatureGeneratorOEBasicHash = SignatureGeneratorOEBasicHash def find_siginfo(pn, taskname, taskhashlist, d): """ Find signature data files for comparison purposes """ import fnmatch import glob if taskhashlist: hashfiles = {} if not taskname: # We have to derive pn and taskname key = pn splitit = key.split('.bb.') taskname = splitit[1] pn = os.path.basename(splitit[0]).split('_')[0] if key.startswith('virtual:native:'): pn = pn + '-native' if taskname in ['do_fetch', 'do_unpack', 'do_patch', 'do_populate_lic']: pn.replace("-native", "") filedates = {} # First search in stamps dir localdata = d.createCopy() localdata.setVar('MULTIMACH_TARGET_SYS', '*') localdata.setVar('PN', pn) localdata.setVar('PV', '*') localdata.setVar('PR', '*') localdata.setVar('EXTENDPE', '') stamp = localdata.getVar('STAMP', True) filespec = '%s.%s.sigdata.*' % (stamp, taskname) foundall = False import glob for fullpath in glob.glob(filespec): match = False if taskhashlist: for taskhash in taskhashlist: if fullpath.endswith('.%s' % taskhash): hashfiles[taskhash] = fullpath if len(hashfiles) == len(taskhashlist): foundall = True break else: filedates[fullpath] = os.stat(fullpath).st_mtime if not taskhashlist or (len(filedates) < 2 and not foundall): # That didn't work, look in sstate-cache hashes = taskhashlist or ['*'] localdata = bb.data.createCopy(d) for hashval in hashes: localdata.setVar('PACKAGE_ARCH', '*') localdata.setVar('TARGET_VENDOR', '*') localdata.setVar('TARGET_OS', '*') localdata.setVar('PN', pn) localdata.setVar('PV', '*') localdata.setVar('PR', '*') localdata.setVar('BB_TASKHASH', hashval) if pn.endswith('-native') or "-cross-" in pn or "-crosssdk-" in pn: localdata.setVar('SSTATE_EXTRAPATH', "${NATIVELSBSTRING}/") sstatename = taskname[3:] filespec = '%s_%s.*.siginfo' % (localdata.getVar('SSTATE_PKG', True), sstatename) if hashval != '*': sstatedir = "%s/%s" % (d.getVar('SSTATE_DIR', True), hashval[:2]) else: sstatedir = d.getVar('SSTATE_DIR', True) for root, dirs, files in os.walk(sstatedir): for fn in files: fullpath = os.path.join(root, fn) if fnmatch.fnmatch(fullpath, filespec): if taskhashlist: hashfiles[hashval] = fullpath else: filedates[fullpath] = os.stat(fullpath).st_mtime if taskhashlist: return hashfiles else: return filedates bb.siggen.find_siginfo = find_siginfo ########NEW FILE######## __FILENAME__ = terminal import logging import oe.classutils import shlex from bb.process import Popen, ExecutionError logger = logging.getLogger('BitBake.OE.Terminal') class UnsupportedTerminal(Exception): pass class NoSupportedTerminals(Exception): pass class Registry(oe.classutils.ClassRegistry): command = None def __init__(cls, name, bases, attrs): super(Registry, cls).__init__(name.lower(), bases, attrs) @property def implemented(cls): return bool(cls.command) class Terminal(Popen): __metaclass__ = Registry def __init__(self, sh_cmd, title=None, env=None, d=None): fmt_sh_cmd = self.format_command(sh_cmd, title) try: Popen.__init__(self, fmt_sh_cmd, env=env) except OSError as exc: import errno if exc.errno == errno.ENOENT: raise UnsupportedTerminal(self.name) else: raise def format_command(self, sh_cmd, title): fmt = {'title': title or 'Terminal', 'command': sh_cmd} if isinstance(self.command, basestring): return shlex.split(self.command.format(**fmt)) else: return [element.format(**fmt) for element in self.command] class XTerminal(Terminal): def __init__(self, sh_cmd, title=None, env=None, d=None): Terminal.__init__(self, sh_cmd, title, env, d) if not os.environ.get('DISPLAY'): raise UnsupportedTerminal(self.name) class Gnome(XTerminal): command = 'gnome-terminal -t "{title}" -x {command}' priority = 2 class Mate(XTerminal): command = 'mate-terminal -t "{title}" -x {command}' priority = 2 class Xfce(XTerminal): command = 'xfce4-terminal -T "{title}" -e "{command}"' priority = 2 class Konsole(XTerminal): command = 'konsole -T "{title}" -e {command}' priority = 2 def __init__(self, sh_cmd, title=None, env=None, d=None): # Check version vernum = check_konsole_version("konsole") if vernum: if vernum.split('.')[0] == "2": logger.debug(1, 'Konsole from KDE 4.x will not work as devshell, skipping') raise UnsupportedTerminal(self.name) XTerminal.__init__(self, sh_cmd, title, env, d) class XTerm(XTerminal): command = 'xterm -T "{title}" -e {command}' priority = 1 class Rxvt(XTerminal): command = 'rxvt -T "{title}" -e {command}' priority = 1 class Screen(Terminal): command = 'screen -D -m -t "{title}" -S devshell {command}' def __init__(self, sh_cmd, title=None, env=None, d=None): s_id = "devshell_%i" % os.getpid() self.command = "screen -D -m -t \"{title}\" -S %s {command}" % s_id Terminal.__init__(self, sh_cmd, title, env, d) msg = 'Screen started. Please connect in another terminal with ' \ '"screen -r %s"' % s_id if (d): bb.event.fire(bb.event.LogExecTTY(msg, "screen -r %s" % s_id, 0.5, 10), d) else: logger.warn(msg) class TmuxRunning(Terminal): """Open a new pane in the current running tmux window""" name = 'tmux-running' command = 'tmux split-window "{command}"' priority = 2.75 def __init__(self, sh_cmd, title=None, env=None, d=None): if not bb.utils.which(os.getenv('PATH'), 'tmux'): raise UnsupportedTerminal('tmux is not installed') if not os.getenv('TMUX'): raise UnsupportedTerminal('tmux is not running') Terminal.__init__(self, sh_cmd, title, env, d) class Tmux(Terminal): """Start a new tmux session and window""" command = 'tmux new -d -s devshell -n devshell "{command}"' priority = 0.75 def __init__(self, sh_cmd, title=None, env=None, d=None): if not bb.utils.which(os.getenv('PATH'), 'tmux'): raise UnsupportedTerminal('tmux is not installed') # TODO: consider using a 'devshell' session shared amongst all # devshells, if it's already there, add a new window to it. window_name = 'devshell-%i' % os.getpid() self.command = 'tmux new -d -s {0} -n {0} "{{command}}"'.format(window_name) Terminal.__init__(self, sh_cmd, title, env, d) attach_cmd = 'tmux att -t {0}'.format(window_name) msg = 'Tmux started. Please connect in another terminal with `tmux att -t {0}`'.format(window_name) if d: bb.event.fire(bb.event.LogExecTTY(msg, attach_cmd, 0.5, 10), d) else: logger.warn(msg) class Custom(Terminal): command = 'false' # This is a placeholder priority = 3 def __init__(self, sh_cmd, title=None, env=None, d=None): self.command = d and d.getVar('OE_TERMINAL_CUSTOMCMD', True) if self.command: if not '{command}' in self.command: self.command += ' {command}' Terminal.__init__(self, sh_cmd, title, env, d) logger.warn('Custom terminal was started.') else: logger.debug(1, 'No custom terminal (OE_TERMINAL_CUSTOMCMD) set') raise UnsupportedTerminal('OE_TERMINAL_CUSTOMCMD not set') def prioritized(): return Registry.prioritized() def spawn_preferred(sh_cmd, title=None, env=None, d=None): """Spawn the first supported terminal, by priority""" for terminal in prioritized(): try: spawn(terminal.name, sh_cmd, title, env, d) break except UnsupportedTerminal: continue else: raise NoSupportedTerminals() def spawn(name, sh_cmd, title=None, env=None, d=None): """Spawn the specified terminal, by name""" logger.debug(1, 'Attempting to spawn terminal "%s"', name) try: terminal = Registry.registry[name] except KeyError: raise UnsupportedTerminal(name) pipe = terminal(sh_cmd, title, env, d) output = pipe.communicate()[0] if pipe.returncode != 0: raise ExecutionError(sh_cmd, pipe.returncode, output) def check_konsole_version(konsole): import subprocess as sub try: p = sub.Popen(['sh', '-c', '%s --version' % konsole],stdout=sub.PIPE,stderr=sub.PIPE) out, err = p.communicate() ver_info = out.rstrip().split('\n') except OSError as exc: import errno if exc.errno == errno.ENOENT: return None else: raise vernum = None for ver in ver_info: if ver.startswith('Konsole'): vernum = ver.split(' ')[-1] return vernum def distro_name(): try: p = Popen(['lsb_release', '-i']) out, err = p.communicate() distro = out.split(':')[1].strip().lower() except: distro = "unknown" return distro ########NEW FILE######## __FILENAME__ = test_license import unittest import oe.license class SeenVisitor(oe.license.LicenseVisitor): def __init__(self): self.seen = [] oe.license.LicenseVisitor.__init__(self) def visit_Str(self, node): self.seen.append(node.s) class TestSingleLicense(unittest.TestCase): licenses = [ "GPLv2", "LGPL-2.0", "Artistic", "MIT", "GPLv3+", "FOO_BAR", ] invalid_licenses = ["GPL/BSD"] @staticmethod def parse(licensestr): visitor = SeenVisitor() visitor.visit_string(licensestr) return visitor.seen def test_single_licenses(self): for license in self.licenses: licenses = self.parse(license) self.assertListEqual(licenses, [license]) def test_invalid_licenses(self): for license in self.invalid_licenses: with self.assertRaises(oe.license.InvalidLicense) as cm: self.parse(license) self.assertEqual(cm.exception.license, license) class TestSimpleCombinations(unittest.TestCase): tests = { "FOO&BAR": ["FOO", "BAR"], "BAZ & MOO": ["BAZ", "MOO"], "ALPHA|BETA": ["ALPHA"], "BAZ&MOO|FOO": ["FOO"], "FOO&BAR|BAZ": ["FOO", "BAR"], } preferred = ["ALPHA", "FOO", "BAR"] def test_tests(self): def choose(a, b): if all(lic in self.preferred for lic in b): return b else: return a for license, expected in self.tests.items(): licenses = oe.license.flattened_licenses(license, choose) self.assertListEqual(licenses, expected) class TestComplexCombinations(TestSimpleCombinations): tests = { "FOO & (BAR | BAZ)&MOO": ["FOO", "BAR", "MOO"], "(ALPHA|(BETA&THETA)|OMEGA)&DELTA": ["OMEGA", "DELTA"], "((ALPHA|BETA)&FOO)|BAZ": ["BETA", "FOO"], "(GPL-2.0|Proprietary)&BSD-4-clause&MIT": ["GPL-2.0", "BSD-4-clause", "MIT"], } preferred = ["BAR", "OMEGA", "BETA", "GPL-2.0"] ########NEW FILE######## __FILENAME__ = test_path import unittest import oe, oe.path import tempfile import os import errno import shutil class TestRealPath(unittest.TestCase): DIRS = [ "a", "b", "etc", "sbin", "usr", "usr/bin", "usr/binX", "usr/sbin", "usr/include", "usr/include/gdbm" ] FILES = [ "etc/passwd", "b/file" ] LINKS = [ ( "bin", "/usr/bin", "/usr/bin" ), ( "binX", "usr/binX", "/usr/binX" ), ( "c", "broken", "/broken" ), ( "etc/passwd-1", "passwd", "/etc/passwd" ), ( "etc/passwd-2", "passwd-1", "/etc/passwd" ), ( "etc/passwd-3", "/etc/passwd-1", "/etc/passwd" ), ( "etc/shadow-1", "/etc/shadow", "/etc/shadow" ), ( "etc/shadow-2", "/etc/shadow-1", "/etc/shadow" ), ( "prog-A", "bin/prog-A", "/usr/bin/prog-A" ), ( "prog-B", "/bin/prog-B", "/usr/bin/prog-B" ), ( "usr/bin/prog-C", "../../sbin/prog-C", "/sbin/prog-C" ), ( "usr/bin/prog-D", "/sbin/prog-D", "/sbin/prog-D" ), ( "usr/binX/prog-E", "../sbin/prog-E", None ), ( "usr/bin/prog-F", "../../../sbin/prog-F", "/sbin/prog-F" ), ( "loop", "a/loop", None ), ( "a/loop", "../loop", None ), ( "b/test", "file/foo", "/b/file/foo" ), ] LINKS_PHYS = [ ( "./", "/", "" ), ( "binX/prog-E", "/usr/sbin/prog-E", "/sbin/prog-E" ), ] EXCEPTIONS = [ ( "loop", errno.ELOOP ), ( "b/test", errno.ENOENT ), ] def __del__(self): try: #os.system("tree -F %s" % self.tmpdir) shutil.rmtree(self.tmpdir) except: pass def setUp(self): self.tmpdir = tempfile.mkdtemp(prefix = "oe-test_path") self.root = os.path.join(self.tmpdir, "R") os.mkdir(os.path.join(self.tmpdir, "_real")) os.symlink("_real", self.root) for d in self.DIRS: os.mkdir(os.path.join(self.root, d)) for f in self.FILES: file(os.path.join(self.root, f), "w") for l in self.LINKS: os.symlink(l[1], os.path.join(self.root, l[0])) def __realpath(self, file, use_physdir, assume_dir = True): return oe.path.realpath(os.path.join(self.root, file), self.root, use_physdir, assume_dir = assume_dir) def test_norm(self): for l in self.LINKS: if l[2] == None: continue target_p = self.__realpath(l[0], True) target_l = self.__realpath(l[0], False) if l[2] != False: self.assertEqual(target_p, target_l) self.assertEqual(l[2], target_p[len(self.root):]) def test_phys(self): for l in self.LINKS_PHYS: target_p = self.__realpath(l[0], True) target_l = self.__realpath(l[0], False) self.assertEqual(l[1], target_p[len(self.root):]) self.assertEqual(l[2], target_l[len(self.root):]) def test_loop(self): for e in self.EXCEPTIONS: self.assertRaisesRegexp(OSError, r'\[Errno %u\]' % e[1], self.__realpath, e[0], False, False) ########NEW FILE######## __FILENAME__ = test_types import unittest from oe.maketype import create, factory class TestTypes(unittest.TestCase): def assertIsInstance(self, obj, cls): return self.assertTrue(isinstance(obj, cls)) def assertIsNot(self, obj, other): return self.assertFalse(obj is other) def assertFactoryCreated(self, value, type, **flags): cls = factory(type) self.assertIsNot(cls, None) self.assertIsInstance(create(value, type, **flags), cls) class TestBooleanType(TestTypes): def test_invalid(self): self.assertRaises(ValueError, create, '', 'boolean') self.assertRaises(ValueError, create, 'foo', 'boolean') self.assertRaises(TypeError, create, object(), 'boolean') def test_true(self): self.assertTrue(create('y', 'boolean')) self.assertTrue(create('yes', 'boolean')) self.assertTrue(create('1', 'boolean')) self.assertTrue(create('t', 'boolean')) self.assertTrue(create('true', 'boolean')) self.assertTrue(create('TRUE', 'boolean')) self.assertTrue(create('truE', 'boolean')) def test_false(self): self.assertFalse(create('n', 'boolean')) self.assertFalse(create('no', 'boolean')) self.assertFalse(create('0', 'boolean')) self.assertFalse(create('f', 'boolean')) self.assertFalse(create('false', 'boolean')) self.assertFalse(create('FALSE', 'boolean')) self.assertFalse(create('faLse', 'boolean')) def test_bool_equality(self): self.assertEqual(create('n', 'boolean'), False) self.assertNotEqual(create('n', 'boolean'), True) self.assertEqual(create('y', 'boolean'), True) self.assertNotEqual(create('y', 'boolean'), False) class TestList(TestTypes): def assertListEqual(self, value, valid, sep=None): obj = create(value, 'list', separator=sep) self.assertEqual(obj, valid) if sep is not None: self.assertEqual(obj.separator, sep) self.assertEqual(str(obj), obj.separator.join(obj)) def test_list_nosep(self): testlist = ['alpha', 'beta', 'theta'] self.assertListEqual('alpha beta theta', testlist) self.assertListEqual('alpha beta\ttheta', testlist) self.assertListEqual('alpha', ['alpha']) def test_list_usersep(self): self.assertListEqual('foo:bar', ['foo', 'bar'], ':') self.assertListEqual('foo:bar:baz', ['foo', 'bar', 'baz'], ':') ########NEW FILE######## __FILENAME__ = test_utils import unittest from oe.utils import packages_filter_out_system class TestPackagesFilterOutSystem(unittest.TestCase): def test_filter(self): """ Test that oe.utils.packages_filter_out_system works. """ try: import bb except ImportError: self.skipTest("Cannot import bb") d = bb.data_smart.DataSmart() d.setVar("PN", "foo") d.setVar("PACKAGES", "foo foo-doc foo-dev") pkgs = packages_filter_out_system(d) self.assertEqual(pkgs, []) d.setVar("PACKAGES", "foo foo-doc foo-data foo-dev") pkgs = packages_filter_out_system(d) self.assertEqual(pkgs, ["foo-data"]) d.setVar("PACKAGES", "foo foo-locale-en-gb") pkgs = packages_filter_out_system(d) self.assertEqual(pkgs, []) d.setVar("PACKAGES", "foo foo-data foo-locale-en-gb") pkgs = packages_filter_out_system(d) self.assertEqual(pkgs, ["foo-data"]) class TestTrimVersion(unittest.TestCase): def test_version_exception(self): with self.assertRaises(TypeError): trim_version(None, 2) with self.assertRaises(TypeError): trim_version((1, 2, 3), 2) def test_num_exception(self): with self.assertRaises(ValueError): trim_version("1.2.3", 0) with self.assertRaises(ValueError): trim_version("1.2.3", -1) def test_valid(self): self.assertEqual(trim_version("1.2.3", 1), "1") self.assertEqual(trim_version("1.2.3", 2), "1.2") self.assertEqual(trim_version("1.2.3", 3), "1.2.3") self.assertEqual(trim_version("1.2.3", 4), "1.2.3") ########NEW FILE######## __FILENAME__ = types import errno import re import os class OEList(list): """OpenEmbedded 'list' type Acts as an ordinary list, but is constructed from a string value and a separator (optional), and re-joins itself when converted to a string with str(). Set the variable type flag to 'list' to use this type, and the 'separator' flag may be specified (defaulting to whitespace).""" name = "list" def __init__(self, value, separator = None): if value is not None: list.__init__(self, value.split(separator)) else: list.__init__(self) if separator is None: self.separator = " " else: self.separator = separator def __str__(self): return self.separator.join(self) def choice(value, choices): """OpenEmbedded 'choice' type Acts as a multiple choice for the user. To use this, set the variable type flag to 'choice', and set the 'choices' flag to a space separated list of valid values.""" if not isinstance(value, basestring): raise TypeError("choice accepts a string, not '%s'" % type(value)) value = value.lower() choices = choices.lower() if value not in choices.split(): raise ValueError("Invalid choice '%s'. Valid choices: %s" % (value, choices)) return value class NoMatch(object): """Stub python regex pattern object which never matches anything""" def findall(self, string, flags=0): return None def finditer(self, string, flags=0): return None def match(self, flags=0): return None def search(self, string, flags=0): return None def split(self, string, maxsplit=0): return None def sub(pattern, repl, string, count=0): return None def subn(pattern, repl, string, count=0): return None NoMatch = NoMatch() def regex(value, regexflags=None): """OpenEmbedded 'regex' type Acts as a regular expression, returning the pre-compiled regular expression pattern object. To use this type, set the variable type flag to 'regex', and optionally, set the 'regexflags' type to a space separated list of the flags to control the regular expression matching (e.g. FOO[regexflags] += 'ignorecase'). See the python documentation on the 're' module for a list of valid flags.""" flagval = 0 if regexflags: for flag in regexflags.split(): flag = flag.upper() try: flagval |= getattr(re, flag) except AttributeError: raise ValueError("Invalid regex flag '%s'" % flag) if not value: # Let's ensure that the default behavior for an undefined or empty # variable is to match nothing. If the user explicitly wants to match # anything, they can match '.*' instead. return NoMatch try: return re.compile(value, flagval) except re.error as exc: raise ValueError("Invalid regex value '%s': %s" % (value, exc.args[0])) def boolean(value): """OpenEmbedded 'boolean' type Valid values for true: 'yes', 'y', 'true', 't', '1' Valid values for false: 'no', 'n', 'false', 'f', '0' """ if not isinstance(value, basestring): raise TypeError("boolean accepts a string, not '%s'" % type(value)) value = value.lower() if value in ('yes', 'y', 'true', 't', '1'): return True elif value in ('no', 'n', 'false', 'f', '0'): return False raise ValueError("Invalid boolean value '%s'" % value) def integer(value, numberbase=10): """OpenEmbedded 'integer' type Defaults to base 10, but this can be specified using the optional 'numberbase' flag.""" return int(value, int(numberbase)) _float = float def float(value, fromhex='false'): """OpenEmbedded floating point type To use this type, set the type flag to 'float', and optionally set the 'fromhex' flag to a true value (obeying the same rules as for the 'boolean' type) if the value is in base 16 rather than base 10.""" if boolean(fromhex): return _float.fromhex(value) else: return _float(value) def path(value, relativeto='', normalize='true', mustexist='false'): value = os.path.join(relativeto, value) if boolean(normalize): value = os.path.normpath(value) if boolean(mustexist): try: open(value, 'r') except IOError as exc: if exc.errno == errno.ENOENT: raise ValueError("{0}: {1}".format(value, os.strerror(errno.ENOENT))) return value ########NEW FILE######## __FILENAME__ = utils try: # Python 2 import commands as cmdstatus except ImportError: # Python 3 import subprocess as cmdstatus def read_file(filename): try: f = open( filename, "r" ) except IOError as reason: return "" # WARNING: can't raise an error now because of the new RDEPENDS handling. This is a bit ugly. :M: else: data = f.read().strip() f.close() return data return None def ifelse(condition, iftrue = True, iffalse = False): if condition: return iftrue else: return iffalse def conditional(variable, checkvalue, truevalue, falsevalue, d): if d.getVar(variable,1) == checkvalue: return truevalue else: return falsevalue def less_or_equal(variable, checkvalue, truevalue, falsevalue, d): if float(d.getVar(variable,1)) <= float(checkvalue): return truevalue else: return falsevalue def version_less_or_equal(variable, checkvalue, truevalue, falsevalue, d): result = bb.utils.vercmp_string(d.getVar(variable,True), checkvalue) if result <= 0: return truevalue else: return falsevalue def both_contain(variable1, variable2, checkvalue, d): if d.getVar(variable1,1).find(checkvalue) != -1 and d.getVar(variable2,1).find(checkvalue) != -1: return checkvalue else: return "" def prune_suffix(var, suffixes, d): # See if var ends with any of the suffixes listed and # remove it if found for suffix in suffixes: if var.endswith(suffix): var = var.replace(suffix, "") prefix = d.getVar("MLPREFIX", True) if prefix and var.startswith(prefix): var = var.replace(prefix, "") return var def str_filter(f, str, d): from re import match return " ".join(filter(lambda x: match(f, x, 0), str.split())) def str_filter_out(f, str, d): from re import match return " ".join(filter(lambda x: not match(f, x, 0), str.split())) def param_bool(cfg, field, dflt = None): """Lookup <field> in <cfg> map and convert it to a boolean; take <dflt> when this <field> does not exist""" value = cfg.get(field, dflt) strvalue = str(value).lower() if strvalue in ('yes', 'y', 'true', 't', '1'): return True elif strvalue in ('no', 'n', 'false', 'f', '0'): return False raise ValueError("invalid value for boolean parameter '%s': '%s'" % (field, value)) def inherits(d, *classes): """Return True if the metadata inherits any of the specified classes""" return any(bb.data.inherits_class(cls, d) for cls in classes) def features_backfill(var,d): # This construct allows the addition of new features to variable specified # as var # Example for var = "DISTRO_FEATURES" # This construct allows the addition of new features to DISTRO_FEATURES # that if not present would disable existing functionality, without # disturbing distributions that have already set DISTRO_FEATURES. # Distributions wanting to elide a value in DISTRO_FEATURES_BACKFILL should # add the feature to DISTRO_FEATURES_BACKFILL_CONSIDERED features = (d.getVar(var, True) or "").split() backfill = (d.getVar(var+"_BACKFILL", True) or "").split() considered = (d.getVar(var+"_BACKFILL_CONSIDERED", True) or "").split() addfeatures = [] for feature in backfill: if feature not in features and feature not in considered: addfeatures.append(feature) if addfeatures: d.appendVar(var, " " + " ".join(addfeatures)) def packages_filter_out_system(d): """ Return a list of packages from PACKAGES with the "system" packages such as PN-dbg PN-doc PN-locale-eb-gb removed. """ pn = d.getVar('PN', True) blacklist = map(lambda suffix: pn + suffix, ('', '-dbg', '-dev', '-doc', '-locale', '-staticdev')) localepkg = pn + "-locale-" pkgs = [] for pkg in d.getVar('PACKAGES', True).split(): if pkg not in blacklist and localepkg not in pkg: pkgs.append(pkg) return pkgs def getstatusoutput(cmd): return cmdstatus.getstatusoutput(cmd) def trim_version(version, num_parts=2): """ Return just the first <num_parts> of <version>, split by periods. For example, trim_version("1.2.3", 2) will return "1.2". """ if type(version) is not str: raise TypeError("Version should be a string") if num_parts < 1: raise ValueError("Cannot split to parts < 1") parts = version.split(".") trimmed = ".".join(parts[:num_parts]) return trimmed def cpu_count(): import multiprocessing return multiprocessing.cpu_count() def execute_pre_post_process(d, cmds): if cmds is None: return for cmd in cmds.strip().split(';'): cmd = cmd.strip() if cmd != '': bb.note("Executing %s ..." % cmd) bb.build.exec_func(cmd, d) ########NEW FILE######## __FILENAME__ = masterimage # Copyright (C) 2014 Intel Corporation # # Released under the MIT license (see COPYING.MIT) # This module adds support to testimage.bbclass to deploy images and run # tests using a "master image" - this is a "known good" image that is # installed onto the device as part of initial setup and will be booted into # with no interaction; we can then use it to deploy the image to be tested # to a second partition before running the tests. # # For an example master image, see core-image-testmaster # (meta/recipes-extended/images/core-image-testmaster.bb) import os import bb import traceback import time import subprocess import oeqa.targetcontrol import oeqa.utils.sshcontrol as sshcontrol import oeqa.utils.commands as commands from oeqa.utils import CommandError from abc import ABCMeta, abstractmethod class MasterImageHardwareTarget(oeqa.targetcontrol.BaseTarget): __metaclass__ = ABCMeta def __init__(self, d): super(MasterImageHardwareTarget, self).__init__(d) # target ip addr = d.getVar("TEST_TARGET_IP", True) or bb.fatal('Please set TEST_TARGET_IP with the IP address of the machine you want to run the tests on.') self.ip = addr.split(":")[0] try: self.port = addr.split(":")[1] except IndexError: self.port = None bb.note("Target IP: %s" % self.ip) self.server_ip = d.getVar("TEST_SERVER_IP", True) if not self.server_ip: try: self.server_ip = subprocess.check_output(['ip', 'route', 'get', self.ip ]).split("\n")[0].split()[-1] except Exception as e: bb.fatal("Failed to determine the host IP address (alternatively you can set TEST_SERVER_IP with the IP address of this machine): %s" % e) bb.note("Server IP: %s" % self.server_ip) # test rootfs + kernel self.rootfs = os.path.join(d.getVar("DEPLOY_DIR_IMAGE", True), d.getVar("IMAGE_LINK_NAME", True) + '.tar.gz') self.kernel = os.path.join(d.getVar("DEPLOY_DIR_IMAGE", True), d.getVar("KERNEL_IMAGETYPE")) if not os.path.isfile(self.rootfs): # we could've checked that IMAGE_FSTYPES contains tar.gz but the config for running testimage might not be # the same as the config with which the image was build, ie # you bitbake core-image-sato with IMAGE_FSTYPES += "tar.gz" # and your autobuilder overwrites the config, adds the test bits and runs bitbake core-image-sato -c testimage bb.fatal("No rootfs found. Did you build the image ?\nIf yes, did you build it with IMAGE_FSTYPES += \"tar.gz\" ? \ \nExpected path: %s" % self.rootfs) if not os.path.isfile(self.kernel): bb.fatal("No kernel found. Expected path: %s" % self.kernel) # master ssh connection self.master = None # if the user knows what they are doing, then by all means... self.user_cmds = d.getVar("TEST_DEPLOY_CMDS", True) self.deploy_cmds = None # this is the name of the command that controls the power for a board # e.g: TEST_POWERCONTROL_CMD = "/home/user/myscripts/powercontrol.py ${MACHINE} what-ever-other-args-the-script-wants" # the command should take as the last argument "off" and "on" and "cycle" (off, on) self.powercontrol_cmd = d.getVar("TEST_POWERCONTROL_CMD", True) or None self.powercontrol_args = d.getVar("TEST_POWERCONTROL_EXTRA_ARGS") or "" self.serialcontrol_cmd = d.getVar("TEST_SERIALCONTROL_CMD", True) or None self.serialcontrol_args = d.getVar("TEST_SERIALCONTROL_EXTRA_ARGS") or "" self.origenv = os.environ if self.powercontrol_cmd or self.serialcontrol_cmd: # the external script for controlling power might use ssh # ssh + keys means we need the original user env bborigenv = d.getVar("BB_ORIGENV", False) or {} for key in bborigenv: val = bborigenv.getVar(key, True) if val is not None: self.origenv[key] = str(val) if self.powercontrol_cmd: if self.powercontrol_args: self.powercontrol_cmd = "%s %s" % (self.powercontrol_cmd, self.powercontrol_args) if self.serialcontrol_cmd: if self.serialcontrol_args: self.serialcontrol_cmd = "%s %s" % (self.serialcontrol_cmd, self.serialcontrol_args) def power_ctl(self, msg): if self.powercontrol_cmd: cmd = "%s %s" % (self.powercontrol_cmd, msg) try: commands.runCmd(cmd, assert_error=False, preexec_fn=os.setsid, env=self.origenv) except CommandError as e: bb.fatal(str(e)) def power_cycle(self, conn): if self.powercontrol_cmd: # be nice, don't just cut power conn.run("shutdown -h now") time.sleep(10) self.power_ctl("cycle") else: status, output = conn.run("reboot") if status != 0: bb.error("Failed rebooting target and no power control command defined. You need to manually reset the device.\n%s" % output) def _wait_until_booted(self): ''' Waits until the target device has booted (if we have just power cycled it) ''' # Subclasses with better methods of determining boot can override this time.sleep(120) def deploy(self): # base class just sets the ssh log file for us super(MasterImageHardwareTarget, self).deploy() self.master = sshcontrol.SSHControl(ip=self.ip, logfile=self.sshlog, timeout=600, port=self.port) status, output = self.master.run("cat /etc/masterimage") if status != 0: # We're not booted into the master image, so try rebooting bb.plain("%s - booting into the master image" % self.pn) self.power_ctl("cycle") self._wait_until_booted() bb.plain("%s - deploying image on target" % self.pn) status, output = self.master.run("cat /etc/masterimage") if status != 0: bb.fatal("No ssh connectivity or target isn't running a master image.\n%s" % output) if self.user_cmds: self.deploy_cmds = self.user_cmds.split("\n") try: self._deploy() except Exception as e: bb.fatal("Failed deploying test image: %s" % e) @abstractmethod def _deploy(self): pass def start(self, params=None): bb.plain("%s - boot test image on target" % self.pn) self._start() # set the ssh object for the target/test image self.connection = sshcontrol.SSHControl(self.ip, logfile=self.sshlog, port=self.port) bb.plain("%s - start running tests" % self.pn) @abstractmethod def _start(self): pass def stop(self): bb.plain("%s - reboot/powercycle target" % self.pn) self.power_cycle(self.connection) class GummibootTarget(MasterImageHardwareTarget): def __init__(self, d): super(GummibootTarget, self).__init__(d) # this the value we need to set in the LoaderEntryOneShot EFI variable # so the system boots the 'test' bootloader label and not the default # The first four bytes are EFI bits, and the rest is an utf-16le string # (EFI vars values need to be utf-16) # $ echo -en "test\0" | iconv -f ascii -t utf-16le | hexdump -C # 00000000 74 00 65 00 73 00 74 00 00 00 |t.e.s.t...| self.efivarvalue = r'\x07\x00\x00\x00\x74\x00\x65\x00\x73\x00\x74\x00\x00\x00' self.deploy_cmds = [ 'mount -L boot /boot', 'mkdir -p /mnt/testrootfs', 'mount -L testrootfs /mnt/testrootfs', 'modprobe efivarfs', 'mount -t efivarfs efivarfs /sys/firmware/efi/efivars', 'cp ~/test-kernel /boot', 'rm -rf /mnt/testrootfs/*', 'tar xzvf ~/test-rootfs.tar.gz -C /mnt/testrootfs', 'printf "%s" > /sys/firmware/efi/efivars/LoaderEntryOneShot-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f' % self.efivarvalue ] def _deploy(self): # make sure these aren't mounted self.master.run("umount /boot; umount /mnt/testrootfs; umount /sys/firmware/efi/efivars;") # from now on, every deploy cmd should return 0 # else an exception will be thrown by sshcontrol self.master.ignore_status = False self.master.copy_to(self.rootfs, "~/test-rootfs.tar.gz") self.master.copy_to(self.kernel, "~/test-kernel") for cmd in self.deploy_cmds: self.master.run(cmd) def _start(self, params=None): self.power_cycle(self.master) # there are better ways than a timeout but this should work for now time.sleep(120) ########NEW FILE######## __FILENAME__ = testtargetloader import types import bb # This class is responsible for loading a test target controller class TestTargetLoader: # Search oeqa.controllers module directory for and return a controller # corresponding to the given target name. # AttributeError raised if not found. # ImportError raised if a provided module can not be imported. def get_controller_module(self, target, bbpath): controllerslist = self.get_controller_modulenames(bbpath) bb.note("Available controller modules: %s" % str(controllerslist)) controller = self.load_controller_from_name(target, controllerslist) return controller # Return a list of all python modules in lib/oeqa/controllers for each # layer in bbpath def get_controller_modulenames(self, bbpath): controllerslist = [] def add_controller_list(path): if not os.path.exists(os.path.join(path, '__init__.py')): bb.fatal('Controllers directory %s exists but is missing __init__.py' % path) files = sorted([f for f in os.listdir(path) if f.endswith('.py') and not f.startswith('_')]) for f in files: module = 'oeqa.controllers.' + f[:-3] if module not in controllerslist: controllerslist.append(module) else: bb.warn("Duplicate controller module found for %s, only one added. Layers should create unique controller module names" % module) for p in bbpath: controllerpath = os.path.join(p, 'lib', 'oeqa', 'controllers') bb.debug(2, 'Searching for target controllers in %s' % controllerpath) if os.path.exists(controllerpath): add_controller_list(controllerpath) return controllerslist # Search for and return a controller from given target name and # set of module names. # Raise AttributeError if not found. # Raise ImportError if a provided module can not be imported def load_controller_from_name(self, target, modulenames): for name in modulenames: obj = self.load_controller_from_module(target, name) if obj: return obj raise AttributeError("Unable to load {0} from available modules: {1}".format(target, str(modulenames))) # Search for and return a controller or None from given module name def load_controller_from_module(self, target, modulename): obj = None # import module, allowing it to raise import exception module = __import__(modulename, globals(), locals(), [target]) # look for target class in the module, catching any exceptions as it # is valid that a module may not have the target class. try: obj = getattr(module, target) if obj: from oeqa.targetcontrol import BaseTarget if (not isinstance(obj, (type, types.ClassType))): bb.warn("Target {0} found, but not of type Class".format(target)) if( not issubclass(obj, BaseTarget)): bb.warn("Target {0} found, but subclass is not BaseTarget".format(target)) except: obj = None return obj ########NEW FILE######## __FILENAME__ = oetest # Copyright (C) 2013 Intel Corporation # # Released under the MIT license (see COPYING.MIT) # Main unittest module used by testimage.bbclass # This provides the oeRuntimeTest base class which is inherited by all tests in meta/lib/oeqa/runtime. # It also has some helper functions and it's responsible for actually starting the tests import os, re, mmap import unittest import inspect def loadTests(tc): # set the context object passed from the test class setattr(oeTest, "tc", tc) # set ps command to use setattr(oeRuntimeTest, "pscmd", "ps -ef" if oeTest.hasPackage("procps") else "ps") # prepare test suite, loader and runner suite = unittest.TestSuite() testloader = unittest.TestLoader() testloader.sortTestMethodsUsing = None suite = testloader.loadTestsFromNames(tc.testslist) return suite def runTests(tc): suite = loadTests(tc) print("Test modules %s" % tc.testslist) print("Found %s tests" % suite.countTestCases()) runner = unittest.TextTestRunner(verbosity=2) result = runner.run(suite) return result class oeTest(unittest.TestCase): longMessage = True testFailures = [] testSkipped = [] testErrors = [] def run(self, result=None): super(oeTest, self).run(result) # we add to our own lists the results, we use those for decorators if len(result.failures) > len(oeTest.testFailures): oeTest.testFailures.append(str(result.failures[-1][0]).split()[0]) if len(result.skipped) > len(oeTest.testSkipped): oeTest.testSkipped.append(str(result.skipped[-1][0]).split()[0]) if len(result.errors) > len(oeTest.testErrors): oeTest.testErrors.append(str(result.errors[-1][0]).split()[0]) @classmethod def hasPackage(self, pkg): if re.search(pkg, oeTest.tc.pkgmanifest): return True return False @classmethod def hasFeature(self,feature): if feature in oeTest.tc.imagefeatures or \ feature in oeTest.tc.distrofeatures: return True else: return False class oeRuntimeTest(oeTest): def __init__(self, methodName='runTest'): self.target = oeRuntimeTest.tc.target super(oeRuntimeTest, self).__init__(methodName) def getmodule(pos=2): # stack returns a list of tuples containg frame information # First element of the list the is current frame, caller is 1 frameinfo = inspect.stack()[pos] modname = inspect.getmodulename(frameinfo[1]) #modname = inspect.getmodule(frameinfo[0]).__name__ return modname def skipModule(reason, pos=2): modname = getmodule(pos) if modname not in oeTest.tc.testsrequired: raise unittest.SkipTest("%s: %s" % (modname, reason)) else: raise Exception("\nTest %s wants to be skipped.\nReason is: %s" \ "\nTest was required in TEST_SUITES, so either the condition for skipping is wrong" \ "\nor the image really doesn't have the required feature/package when it should." % (modname, reason)) def skipModuleIf(cond, reason): if cond: skipModule(reason, 3) def skipModuleUnless(cond, reason): if not cond: skipModule(reason, 3) ########NEW FILE######## __FILENAME__ = runexported #!/usr/bin/env python # Copyright (C) 2013 Intel Corporation # # Released under the MIT license (see COPYING.MIT) # This script should be used outside of the build system to run image tests. # It needs a json file as input as exported by the build. # E.g for an already built image: #- export the tests: # TEST_EXPORT_ONLY = "1" # TEST_TARGET = "simpleremote" # TEST_TARGET_IP = "192.168.7.2" # TEST_SERVER_IP = "192.168.7.1" # bitbake core-image-sato -c testimage # Setup your target, e.g for qemu: runqemu core-image-sato # cd build/tmp/testimage/core-image-sato # ./runexported.py testdata.json import sys import os import time from optparse import OptionParser try: import simplejson as json except ImportError: import json sys.path.append(os.path.abspath(os.path.join(os.path.dirname(__file__), "oeqa"))) from oeqa.oetest import runTests from oeqa.utils.sshcontrol import SSHControl # this isn't pretty but we need a fake target object # for running the tests externally as we don't care # about deploy/start we only care about the connection methods (run, copy) class FakeTarget(object): def __init__(self, d): self.connection = None self.ip = None self.server_ip = None self.datetime = time.strftime('%Y%m%d%H%M%S',time.gmtime()) self.testdir = d.getVar("TEST_LOG_DIR", True) self.pn = d.getVar("PN", True) def exportStart(self): self.sshlog = os.path.join(self.testdir, "ssh_target_log.%s" % self.datetime) sshloglink = os.path.join(self.testdir, "ssh_target_log") if os.path.islink(sshloglink): os.unlink(sshloglink) os.symlink(self.sshlog, sshloglink) print("SSH log file: %s" % self.sshlog) self.connection = SSHControl(self.ip, logfile=self.sshlog) def run(self, cmd, timeout=None): return self.connection.run(cmd, timeout) def copy_to(self, localpath, remotepath): return self.connection.copy_to(localpath, remotepath) def copy_from(self, remotepath, localpath): return self.connection.copy_from(remotepath, localpath) class MyDataDict(dict): def getVar(self, key, unused = None): return self.get(key, "") class TestContext(object): def __init__(self): self.d = None self.target = None def main(): usage = "usage: %prog [options] <json file>" parser = OptionParser(usage=usage) parser.add_option("-t", "--target-ip", dest="ip", help="The IP address of the target machine. Use this to \ overwrite the value determined from TEST_TARGET_IP at build time") parser.add_option("-s", "--server-ip", dest="server_ip", help="The IP address of this machine. Use this to \ overwrite the value determined from TEST_SERVER_IP at build time.") parser.add_option("-d", "--deploy-dir", dest="deploy_dir", help="Full path to the package feeds, that this \ the contents of what used to be DEPLOY_DIR on the build machine. If not specified it will use the value \ specified in the json if that directory actually exists or it will error out.") parser.add_option("-l", "--log-dir", dest="log_dir", help="This sets the path for TEST_LOG_DIR. If not specified \ the current dir is used. This is used for usually creating a ssh log file and a scp test file.") (options, args) = parser.parse_args() if len(args) != 1: parser.error("Incorrect number of arguments. The one and only argument should be a json file exported by the build system") with open(args[0], "r") as f: loaded = json.load(f) if options.ip: loaded["target"]["ip"] = options.ip if options.server_ip: loaded["target"]["server_ip"] = options.server_ip d = MyDataDict() for key in loaded["d"].keys(): d[key] = loaded["d"][key] if options.log_dir: d["TEST_LOG_DIR"] = options.log_dir else: d["TEST_LOG_DIR"] = os.path.abspath(os.path.dirname(__file__)) if options.deploy_dir: d["DEPLOY_DIR"] = options.deploy_dir else: if not os.path.isdir(d["DEPLOY_DIR"]): raise Exception("The path to DEPLOY_DIR does not exists: %s" % d["DEPLOY_DIR"]) target = FakeTarget(d) for key in loaded["target"].keys(): setattr(target, key, loaded["target"][key]) tc = TestContext() setattr(tc, "d", d) setattr(tc, "target", target) for key in loaded.keys(): if key != "d" and key != "target": setattr(tc, key, loaded[key]) target.exportStart() runTests(tc) return 0 if __name__ == "__main__": try: ret = main() except Exception: ret = 1 import traceback traceback.print_exc(5) sys.exit(ret) ########NEW FILE######## __FILENAME__ = buildcvs from oeqa.oetest import oeRuntimeTest from oeqa.utils.decorators import * from oeqa.utils.targetbuild import TargetBuildProject def setUpModule(): if not oeRuntimeTest.hasFeature("tools-sdk"): skipModule("Image doesn't have tools-sdk in IMAGE_FEATURES") class BuildCvsTest(oeRuntimeTest): @classmethod def setUpClass(self): self.project = TargetBuildProject(oeRuntimeTest.tc.target, oeRuntimeTest.tc.d, "http://ftp.gnu.org/non-gnu/cvs/source/feature/1.12.13/cvs-1.12.13.tar.bz2") self.project.download_archive() @skipUnlessPassed("test_ssh") def test_cvs(self): self.assertEqual(self.project.run_configure(), 0, msg="Running configure failed") self.assertEqual(self.project.run_make(), 0, msg="Running make failed") self.assertEqual(self.project.run_install(), 0, msg="Running make install failed") @classmethod def tearDownClass(self): self.project.clean() ########NEW FILE######## __FILENAME__ = buildiptables from oeqa.oetest import oeRuntimeTest from oeqa.utils.decorators import * from oeqa.utils.targetbuild import TargetBuildProject def setUpModule(): if not oeRuntimeTest.hasFeature("tools-sdk"): skipModule("Image doesn't have tools-sdk in IMAGE_FEATURES") class BuildIptablesTest(oeRuntimeTest): @classmethod def setUpClass(self): self.project = TargetBuildProject(oeRuntimeTest.tc.target, oeRuntimeTest.tc.d, "http://netfilter.org/projects/iptables/files/iptables-1.4.13.tar.bz2") self.project.download_archive() @skipUnlessPassed("test_ssh") def test_iptables(self): self.assertEqual(self.project.run_configure(), 0, msg="Running configure failed") self.assertEqual(self.project.run_make(), 0, msg="Running make failed") self.assertEqual(self.project.run_install(), 0, msg="Running make install failed") @classmethod def tearDownClass(self): self.project.clean() ########NEW FILE######## __FILENAME__ = buildsudoku from oeqa.oetest import oeRuntimeTest from oeqa.utils.decorators import * from oeqa.utils.targetbuild import TargetBuildProject def setUpModule(): if not oeRuntimeTest.hasFeature("tools-sdk"): skipModule("Image doesn't have tools-sdk in IMAGE_FEATURES") class SudokuTest(oeRuntimeTest): @classmethod def setUpClass(self): self.project = TargetBuildProject(oeRuntimeTest.tc.target, oeRuntimeTest.tc.d, "http://downloads.sourceforge.net/project/sudoku-savant/sudoku-savant/sudoku-savant-1.3/sudoku-savant-1.3.tar.bz2") self.project.download_archive() @skipUnlessPassed("test_ssh") def test_sudoku(self): self.assertEqual(self.project.run_configure(), 0, msg="Running configure failed") self.assertEqual(self.project.run_make(), 0, msg="Running make failed") @classmethod def tearDownClass(self): self.project.clean() ########NEW FILE######## __FILENAME__ = connman import unittest from oeqa.oetest import oeRuntimeTest, skipModule from oeqa.utils.decorators import * def setUpModule(): if not oeRuntimeTest.hasPackage("connman"): skipModule("No connman package in image") class ConnmanTest(oeRuntimeTest): def service_status(self, service): if oeRuntimeTest.hasFeature("systemd"): (status, output) = self.target.run('systemctl status -l %s' % service) return output else: return "Unable to get status or logs for %s" % service @skipUnlessPassed('test_ssh') def test_connmand_help(self): (status, output) = self.target.run('/usr/sbin/connmand --help') self.assertEqual(status, 0, msg="status and output: %s and %s" % (status,output)) @skipUnlessPassed('test_connmand_help') def test_connmand_running(self): (status, output) = self.target.run(oeRuntimeTest.pscmd + ' | grep [c]onnmand') if status != 0: print self.service_status("connman") self.fail("No connmand process running") ########NEW FILE######## __FILENAME__ = date from oeqa.oetest import oeRuntimeTest from oeqa.utils.decorators import * import re class DateTest(oeRuntimeTest): @skipUnlessPassed("test_ssh") def test_date(self): (status, output) = self.target.run('date +"%Y-%m-%d %T"') self.assertEqual(status, 0, msg="Failed to get initial date, output: %s" % output) oldDate = output sampleDate = '"2016-08-09 10:00:00"' (status, output) = self.target.run("date -s %s" % sampleDate) self.assertEqual(status, 0, msg="Date set failed, output: %s" % output) (status, output) = self.target.run("date -R") p = re.match('Tue, 09 Aug 2016 10:00:.. \+0000', output) self.assertTrue(p, msg="The date was not set correctly, output: %s" % output) (status, output) = self.target.run('date -s "%s"' % oldDate) self.assertEqual(status, 0, msg="Failed to reset date, output: %s" % output) ########NEW FILE######## __FILENAME__ = df import unittest from oeqa.oetest import oeRuntimeTest from oeqa.utils.decorators import * class DfTest(oeRuntimeTest): @skipUnlessPassed("test_ssh") def test_df(self): (status,output) = self.target.run("df / | sed -n '2p' | awk '{print $4}'") self.assertTrue(int(output)>5120, msg="Not enough space on image. Current size is %s" % output) ########NEW FILE######## __FILENAME__ = dmesg import unittest from oeqa.oetest import oeRuntimeTest from oeqa.utils.decorators import * class DmesgTest(oeRuntimeTest): @skipUnlessPassed('test_ssh') def test_dmesg(self): (status, output) = self.target.run('dmesg | grep -v mmci-pl18x | grep -v "error changing net interface name" | grep -iv "dma timeout" | grep -i error') self.assertEqual(status, 1, msg = "Error messages in dmesg log: %s" % output) ########NEW FILE######## __FILENAME__ = test import os os.system('touch /tmp/testfile.python') a = 9.01e+21 - 9.01e+21 + 0.01 print "the value of a is %s" % a ########NEW FILE######## __FILENAME__ = gcc import unittest import os from oeqa.oetest import oeRuntimeTest, skipModule from oeqa.utils.decorators import * def setUpModule(): if not oeRuntimeTest.hasFeature("tools-sdk"): skipModule("Image doesn't have tools-sdk in IMAGE_FEATURES") class GccCompileTest(oeRuntimeTest): @classmethod def setUpClass(self): oeRuntimeTest.tc.target.copy_to(os.path.join(oeRuntimeTest.tc.filesdir, "test.c"), "/tmp/test.c") oeRuntimeTest.tc.target.copy_to(os.path.join(oeRuntimeTest.tc.filesdir, "testmakefile"), "/tmp/testmakefile") def test_gcc_compile(self): (status, output) = self.target.run('gcc /tmp/test.c -o /tmp/test -lm') self.assertEqual(status, 0, msg="gcc compile failed, output: %s" % output) (status, output) = self.target.run('/tmp/test') self.assertEqual(status, 0, msg="running compiled file failed, output %s" % output) def test_gpp_compile(self): (status, output) = self.target.run('g++ /tmp/test.c -o /tmp/test -lm') self.assertEqual(status, 0, msg="g++ compile failed, output: %s" % output) (status, output) = self.target.run('/tmp/test') self.assertEqual(status, 0, msg="running compiled file failed, output %s" % output) def test_make(self): (status, output) = self.target.run('cd /tmp; make -f testmakefile') self.assertEqual(status, 0, msg="running make failed, output %s" % output) @classmethod def tearDownClass(self): oeRuntimeTest.tc.target.run("rm /tmp/test.c /tmp/test.o /tmp/test /tmp/testmakefile") ########NEW FILE######## __FILENAME__ = kernelmodule import unittest import os from oeqa.oetest import oeRuntimeTest, skipModule from oeqa.utils.decorators import * def setUpModule(): if not oeRuntimeTest.hasFeature("tools-sdk"): skipModule("Image doesn't have tools-sdk in IMAGE_FEATURES") class KernelModuleTest(oeRuntimeTest): def setUp(self): self.target.copy_to(os.path.join(oeRuntimeTest.tc.filesdir, "hellomod.c"), "/tmp/hellomod.c") self.target.copy_to(os.path.join(oeRuntimeTest.tc.filesdir, "hellomod_makefile"), "/tmp/Makefile") @skipUnlessPassed('test_ssh') @skipUnlessPassed('test_gcc_compile') def test_kernel_module(self): cmds = [ 'cd /usr/src/kernel && make scripts', 'cd /tmp && make', 'cd /tmp && insmod hellomod.ko', 'lsmod | grep hellomod', 'dmesg | grep Hello', 'rmmod hellomod', 'dmesg | grep "Cleaning up hellomod"' ] for cmd in cmds: (status, output) = self.target.run(cmd, 900) self.assertEqual(status, 0, msg="\n".join([cmd, output])) def tearDown(self): self.target.run('rm -f /tmp/Makefile /tmp/hellomod.c') ########NEW FILE######## __FILENAME__ = ldd import unittest from oeqa.oetest import oeRuntimeTest from oeqa.utils.decorators import * def setUpModule(): if not oeRuntimeTest.hasFeature("tools-sdk"): skipModule("Image doesn't have tools-sdk in IMAGE_FEATURES") class LddTest(oeRuntimeTest): @skipUnlessPassed('test_ssh') def test_ldd_exists(self): (status, output) = self.target.run('which ldd') self.assertEqual(status, 0, msg = "ldd does not exist in PATH: which ldd: %s" % output) @skipUnlessPassed('test_ldd_exists') def test_ldd_rtldlist_check(self): (status, output) = self.target.run('for i in $(which ldd | xargs cat | grep "^RTLDLIST"|cut -d\'=\' -f2|tr -d \'"\'); do test -f $i && echo $i && break; done') self.assertEqual(status, 0, msg = "ldd path not correct or RTLDLIST files don't exist. ") ########NEW FILE######## __FILENAME__ = logrotate # This test should cover https://bugzilla.yoctoproject.org/tr_show_case.cgi?case_id=289 testcase # Note that the image under test must have logrotate installed import unittest from oeqa.oetest import oeRuntimeTest, skipModule from oeqa.utils.decorators import * def setUpModule(): if not oeRuntimeTest.hasPackage("logrotate"): skipModule("No logrotate package in image") class LogrotateTest(oeRuntimeTest): @skipUnlessPassed("test_ssh") def test_1_logrotate_setup(self): (status, output) = self.target.run('mkdir /home/root/logrotate_dir') self.assertEqual(status, 0, msg = "Could not create logrotate_dir. Output: %s" % output) (status, output) = self.target.run("sed -i 's#wtmp {#wtmp {\\n olddir /home/root/logrotate_dir#' /etc/logrotate.conf") self.assertEqual(status, 0, msg = "Could not write to logrotate.conf file. Status and output: %s and %s)" % (status, output)) @skipUnlessPassed("test_1_logrotate_setup") def test_2_logrotate(self): (status, output) = self.target.run('logrotate -f /etc/logrotate.conf') self.assertEqual(status, 0, msg = "logrotate service could not be reloaded. Status and output: %s and %s" % (status, output)) output = self.target.run('ls -la /home/root/logrotate_dir/ | wc -l')[1] self.assertTrue(int(output)>=3, msg = "new logfile could not be created. List of files within log directory: %s" %(self.target.run('ls -la /home/root/logrotate_dir')[1])) ########NEW FILE######## __FILENAME__ = multilib import unittest from oeqa.oetest import oeRuntimeTest, skipModule from oeqa.utils.decorators import * def setUpModule(): multilibs = oeRuntimeTest.tc.d.getVar("MULTILIBS", True) or "" if "multilib:lib32" not in multilibs: skipModule("this isn't a multilib:lib32 image") class MultilibTest(oeRuntimeTest): @skipUnlessPassed('test_ssh') def test_file_connman(self): self.assertTrue(oeRuntimeTest.hasPackage('connman-gnome'), msg="This test assumes connman-gnome is installed") (status, output) = self.target.run("readelf -h /usr/bin/connman-applet | sed -n '3p' | awk '{print $2}'") self.assertEqual(output, "ELF32", msg="connman-applet isn't an ELF32 binary. readelf says: %s" % self.target.run("readelf -h /usr/bin/connman-applet")[1]) ########NEW FILE######## __FILENAME__ = pam # This test should cover https://bugzilla.yoctoproject.org/tr_show_case.cgi?case_id=287 testcase # Note that the image under test must have "pam" in DISTRO_FEATURES import unittest from oeqa.oetest import oeRuntimeTest from oeqa.utils.decorators import * def setUpModule(): if not oeRuntimeTest.hasFeature("pam"): skipModule("target doesn't have 'pam' in DISTRO_FEATURES") class PamBasicTest(oeRuntimeTest): @skipUnlessPassed('test_ssh') def test_pam(self): (status, output) = self.target.run('login --help') self.assertEqual(status, 1, msg = "login command does not work as expected. Status and output:%s and %s" %(status, output)) (status, output) = self.target.run('passwd --help') self.assertEqual(status, 6, msg = "passwd command does not work as expected. Status and output:%s and %s" %(status, output)) (status, output) = self.target.run('su --help') self.assertEqual(status, 2, msg = "su command does not work as expected. Status and output:%s and %s" %(status, output)) (status, output) = self.target.run('useradd --help') self.assertEqual(status, 2, msg = "useradd command does not work as expected. Status and output:%s and %s" %(status, output)) ########NEW FILE######## __FILENAME__ = perl import unittest import os from oeqa.oetest import oeRuntimeTest, skipModule from oeqa.utils.decorators import * def setUpModule(): if not oeRuntimeTest.hasPackage("perl"): skipModule("No perl package in the image") class PerlTest(oeRuntimeTest): @classmethod def setUpClass(self): oeRuntimeTest.tc.target.copy_to(os.path.join(oeRuntimeTest.tc.filesdir, "test.pl"), "/tmp/test.pl") def test_perl_exists(self): (status, output) = self.target.run('which perl') self.assertEqual(status, 0, msg="Perl binary not in PATH or not on target.") def test_perl_works(self): (status, output) = self.target.run('perl /tmp/test.pl') self.assertEqual(status, 0, msg="Exit status was not 0. Output: %s" % output) self.assertEqual(output, "the value of a is 0.01", msg="Incorrect output: %s" % output) @classmethod def tearDownClass(self): oeRuntimeTest.tc.target.run("rm /tmp/test.pl") ########NEW FILE######## __FILENAME__ = ping import subprocess import unittest import sys import time from oeqa.oetest import oeRuntimeTest class PingTest(oeRuntimeTest): def test_ping(self): output = '' count = 0 endtime = time.time() + 60 while count < 5 and time.time() < endtime: proc = subprocess.Popen("ping -c 1 %s" % self.target.ip, shell=True, stdout=subprocess.PIPE) output += proc.communicate()[0] if proc.poll() == 0: count += 1 else: count = 0 self.assertEqual(count, 5, msg = "Expected 5 consecutive replies, got %d.\nping output is:\n%s" % (count,output)) ########NEW FILE######## __FILENAME__ = python import unittest import os from oeqa.oetest import oeRuntimeTest, skipModule from oeqa.utils.decorators import * def setUpModule(): if not oeRuntimeTest.hasPackage("python"): skipModule("No python package in the image") class PythonTest(oeRuntimeTest): @classmethod def setUpClass(self): oeRuntimeTest.tc.target.copy_to(os.path.join(oeRuntimeTest.tc.filesdir, "test.py"), "/tmp/test.py") def test_python_exists(self): (status, output) = self.target.run('which python') self.assertEqual(status, 0, msg="Python binary not in PATH or not on target.") def test_python_stdout(self): (status, output) = self.target.run('python /tmp/test.py') self.assertEqual(status, 0, msg="Exit status was not 0. Output: %s" % output) self.assertEqual(output, "the value of a is 0.01", msg="Incorrect output: %s" % output) def test_python_testfile(self): (status, output) = self.target.run('ls /tmp/testfile.python') self.assertEqual(status, 0, msg="Python test file generate failed.") @classmethod def tearDownClass(self): oeRuntimeTest.tc.target.run("rm /tmp/test.py /tmp/testfile.python") ########NEW FILE######## __FILENAME__ = rpm import unittest import os import fnmatch from oeqa.oetest import oeRuntimeTest, skipModule from oeqa.utils.decorators import * def setUpModule(): if not oeRuntimeTest.hasFeature("package-management"): skipModule("rpm module skipped: target doesn't have package-management in IMAGE_FEATURES") if "package_rpm" != oeRuntimeTest.tc.d.getVar("PACKAGE_CLASSES", True).split()[0]: skipModule("rpm module skipped: target doesn't have rpm as primary package manager") class RpmBasicTest(oeRuntimeTest): @skipUnlessPassed('test_ssh') def test_rpm_help(self): (status, output) = self.target.run('rpm --help') self.assertEqual(status, 0, msg="status and output: %s and %s" % (status,output)) @skipUnlessPassed('test_rpm_help') def test_rpm_query(self): (status, output) = self.target.run('rpm -q rpm') self.assertEqual(status, 0, msg="status and output: %s and %s" % (status,output)) class RpmInstallRemoveTest(oeRuntimeTest): @classmethod def setUpClass(self): pkgarch = oeRuntimeTest.tc.d.getVar('TUNE_PKGARCH', True).replace("-", "_") rpmdir = os.path.join(oeRuntimeTest.tc.d.getVar('DEPLOY_DIR', True), "rpm", pkgarch) # pick rpm-doc as a test file to get installed, because it's small and it will always be built for standard targets for f in fnmatch.filter(os.listdir(rpmdir), "rpm-doc-*.%s.rpm" % pkgarch): testrpmfile = f oeRuntimeTest.tc.target.copy_to(os.path.join(rpmdir,testrpmfile), "/tmp/rpm-doc.rpm") @skipUnlessPassed('test_rpm_help') def test_rpm_install(self): (status, output) = self.target.run('rpm -ivh /tmp/rpm-doc.rpm') self.assertEqual(status, 0, msg="Failed to install rpm-doc package: %s" % output) @skipUnlessPassed('test_rpm_install') def test_rpm_remove(self): (status,output) = self.target.run('rpm -e rpm-doc') self.assertEqual(status, 0, msg="Failed to remove rpm-doc package: %s" % output) @classmethod def tearDownClass(self): oeRuntimeTest.tc.target.run('rm -f /tmp/rpm-doc.rpm') ########NEW FILE######## __FILENAME__ = scanelf import unittest from oeqa.oetest import oeRuntimeTest, skipModule from oeqa.utils.decorators import * def setUpModule(): if not oeRuntimeTest.hasPackage("pax-utils"): skipModule("pax-utils package not installed") class ScanelfTest(oeRuntimeTest): def setUp(self): self.scancmd = 'scanelf --quiet --recursive --mount --ldpath --path' @skipUnlessPassed('test_ssh') def test_scanelf_textrel(self): # print TEXTREL information self.scancmd += " --textrel" (status, output) = self.target.run(self.scancmd) self.assertEqual(output.strip(), "", "\n".join([self.scancmd, output])) @skipUnlessPassed('test_ssh') def test_scanelf_rpath(self): # print RPATH information self.scancmd += " --rpath" (status, output) = self.target.run(self.scancmd) self.assertEqual(output.strip(), "", "\n".join([self.scancmd, output])) ########NEW FILE######## __FILENAME__ = scp import os from oeqa.oetest import oeRuntimeTest, skipModule from oeqa.utils.decorators import skipUnlessPassed def setUpModule(): if not (oeRuntimeTest.hasPackage("dropbear") or oeRuntimeTest.hasPackage("openssh-sshd")): skipModule("No ssh package in image") class ScpTest(oeRuntimeTest): @skipUnlessPassed('test_ssh') def test_scp_file(self): test_log_dir = oeRuntimeTest.tc.d.getVar("TEST_LOG_DIR", True) test_file_path = os.path.join(test_log_dir, 'test_scp_file') with open(test_file_path, 'w') as test_scp_file: test_scp_file.seek(2 ** 22 - 1) test_scp_file.write(os.linesep) (status, output) = self.target.copy_to(test_file_path, '/tmp/test_scp_file') self.assertEqual(status, 0, msg = "File could not be copied. Output: %s" % output) (status, output) = self.target.run("ls -la /tmp/test_scp_file") self.assertEqual(status, 0, msg = "SCP test failed") ########NEW FILE######## __FILENAME__ = skeletoninit # This test should cover https://bugzilla.yoctoproject.org/tr_show_case.cgi?case_id=284 testcase # Note that the image under test must have meta-skeleton layer in bblayers and IMAGE_INSTALL_append = " service" in local.conf import unittest from oeqa.oetest import oeRuntimeTest from oeqa.utils.decorators import * def setUpModule(): if not oeRuntimeTest.hasPackage("service"): skipModule("No service package in image") class SkeletonBasicTest(oeRuntimeTest): @skipUnlessPassed('test_ssh') @unittest.skipIf("systemd" == oeRuntimeTest.tc.d.getVar("VIRTUAL-RUNTIME_init_manager"), "Not appropiate for systemd image") def test_skeleton_availability(self): (status, output) = self.target.run('ls /etc/init.d/skeleton') self.assertEqual(status, 0, msg = "skeleton init script not found. Output:\n%s " % output) (status, output) = self.target.run('ls /usr/sbin/skeleton-test') self.assertEqual(status, 0, msg = "skeleton-test not found. Output:\n%s" % output) @skipUnlessPassed('test_skeleton_availability') @unittest.skipIf("systemd" == oeRuntimeTest.tc.d.getVar("VIRTUAL-RUNTIME_init_manager"), "Not appropiate for systemd image") def test_skeleton_script(self): output1 = self.target.run("/etc/init.d/skeleton start")[1] (status, output2) = self.target.run(oeRuntimeTest.pscmd + ' | grep [s]keleton-test') self.assertEqual(status, 0, msg = "Skeleton script could not be started:\n%s\n%s" % (output1, output2)) ########NEW FILE######## __FILENAME__ = smart import unittest import re from oeqa.oetest import oeRuntimeTest from oeqa.utils.decorators import * from oeqa.utils.httpserver import HTTPService def setUpModule(): if not oeRuntimeTest.hasFeature("package-management"): skipModule("Image doesn't have package management feature") if not oeRuntimeTest.hasPackage("smart"): skipModule("Image doesn't have smart installed") if "package_rpm" != oeRuntimeTest.tc.d.getVar("PACKAGE_CLASSES", True).split()[0]: skipModule("Rpm is not the primary package manager") class SmartTest(oeRuntimeTest): @skipUnlessPassed('test_smart_help') def smart(self, command, expected = 0): command = 'smart %s' % command status, output = self.target.run(command, 1500) message = os.linesep.join([command, output]) self.assertEqual(status, expected, message) self.assertFalse("Cannot allocate memory" in output, message) return output class SmartBasicTest(SmartTest): @skipUnlessPassed('test_ssh') def test_smart_help(self): self.smart('--help') def test_smart_version(self): self.smart('--version') def test_smart_info(self): self.smart('info python-smartpm') def test_smart_query(self): self.smart('query python-smartpm') def test_smart_search(self): self.smart('search python-smartpm') def test_smart_stats(self): self.smart('stats') class SmartRepoTest(SmartTest): @classmethod def setUpClass(self): self.repo_server = HTTPService(oeRuntimeTest.tc.d.getVar('DEPLOY_DIR', True), oeRuntimeTest.tc.target.server_ip) self.repo_server.start() @classmethod def tearDownClass(self): self.repo_server.stop() def test_smart_channel(self): self.smart('channel', 1) def test_smart_channel_add(self): image_pkgtype = self.tc.d.getVar('IMAGE_PKGTYPE', True) deploy_url = 'http://%s:%s/%s' %(self.target.server_ip, self.repo_server.port, image_pkgtype) pkgarchs = self.tc.d.getVar('PACKAGE_ARCHS', True).replace("-","_").split() for arch in os.listdir('%s/%s' % (self.repo_server.root_dir, image_pkgtype)): if arch in pkgarchs: self.smart('channel -y --add {a} type=rpm-md baseurl={u}/{a}'.format(a=arch, u=deploy_url)) self.smart('update') def test_smart_channel_help(self): self.smart('channel --help') def test_smart_channel_list(self): self.smart('channel --list') def test_smart_channel_show(self): self.smart('channel --show') def test_smart_channel_rpmsys(self): self.smart('channel --show rpmsys') self.smart('channel --disable rpmsys') self.smart('channel --enable rpmsys') @skipUnlessPassed('test_smart_channel_add') def test_smart_install(self): self.smart('remove -y psplash-default') self.smart('install -y psplash-default') @skipUnlessPassed('test_smart_install') def test_smart_install_dependency(self): self.smart('remove -y psplash') self.smart('install -y psplash-default') @skipUnlessPassed('test_smart_channel_add') def test_smart_install_from_disk(self): self.smart('remove -y psplash-default') self.smart('download psplash-default') self.smart('install -y ./psplash-default*') @skipUnlessPassed('test_smart_channel_add') def test_smart_install_from_http(self): output = self.smart('download --urls psplash-default') url = re.search('(http://.*/psplash-default.*\.rpm)', output) self.assertTrue(url, msg="Couln't find download url in %s" % output) self.smart('remove -y psplash-default') self.smart('install -y %s' % url.group(0)) @skipUnlessPassed('test_smart_install') def test_smart_reinstall(self): self.smart('reinstall -y psplash-default') ########NEW FILE######## __FILENAME__ = ssh import subprocess import unittest import sys from oeqa.oetest import oeRuntimeTest, skipModule from oeqa.utils.decorators import * def setUpModule(): if not (oeRuntimeTest.hasPackage("dropbear") or oeRuntimeTest.hasPackage("openssh")): skipModule("No ssh package in image") class SshTest(oeRuntimeTest): @skipUnlessPassed('test_ping') def test_ssh(self): (status, output) = self.target.run('uname -a') self.assertEqual(status, 0, msg="SSH Test failed: %s" % output) (status, output) = self.target.run('cat /etc/masterimage') self.assertEqual(status, 1, msg="This isn't the right image - /etc/masterimage shouldn't be here %s" % output) ########NEW FILE######## __FILENAME__ = syslog import unittest from oeqa.oetest import oeRuntimeTest, skipModule from oeqa.utils.decorators import * def setUpModule(): if not oeRuntimeTest.hasPackage("syslog"): skipModule("No syslog package in image") class SyslogTest(oeRuntimeTest): @skipUnlessPassed("test_ssh") def test_syslog_help(self): (status,output) = self.target.run('/sbin/syslogd --help') self.assertEqual(status, 0, msg="status and output: %s and %s" % (status,output)) @skipUnlessPassed("test_syslog_help") def test_syslog_running(self): (status,output) = self.target.run(oeRuntimeTest.pscmd + ' | grep -i [s]yslogd') self.assertEqual(status, 0, msg="no syslogd process, ps output: %s" % self.target.run(oeRuntimeTest.pscmd)[1]) class SyslogTestConfig(oeRuntimeTest): @skipUnlessPassed("test_syslog_running") def test_syslog_logger(self): (status,output) = self.target.run('logger foobar && test -e /var/log/messages && grep foobar /var/log/messages || logread | grep foobar') self.assertEqual(status, 0, msg="Test log string not found in /var/log/messages. Output: %s " % output) @skipUnlessPassed("test_syslog_running") def test_syslog_restart(self): if "systemd" != oeRuntimeTest.tc.d.getVar("VIRTUAL-RUNTIME_init_manager"): (status,output) = self.target.run('/etc/init.d/syslog restart') else: (status,output) = self.target.run('systemctl restart syslog.service') @skipUnlessPassed("test_syslog_restart") @skipUnlessPassed("test_syslog_logger") @unittest.skipIf("systemd" == oeRuntimeTest.tc.d.getVar("VIRTUAL-RUNTIME_init_manager"), "Not appropiate for systemd image") def test_syslog_startup_config(self): self.target.run('echo "LOGFILE=/var/log/test" >> /etc/syslog-startup.conf') (status,output) = self.target.run('/etc/init.d/syslog restart') self.assertEqual(status, 0, msg="Could not restart syslog service. Status and output: %s and %s" % (status,output)) (status,output) = self.target.run('logger foobar && grep foobar /var/log/test') self.assertEqual(status, 0, msg="Test log string not found. Output: %s " % output) self.target.run("sed -i 's#LOGFILE=/var/log/test##' /etc/syslog-startup.conf") self.target.run('/etc/init.d/syslog restart') ########NEW FILE######## __FILENAME__ = systemd import unittest import re from oeqa.oetest import oeRuntimeTest, skipModule from oeqa.utils.decorators import * def setUpModule(): if not oeRuntimeTest.hasFeature("systemd"): skipModule("target doesn't have systemd in DISTRO_FEATURES") if "systemd" != oeRuntimeTest.tc.d.getVar("VIRTUAL-RUNTIME_init_manager", True): skipModule("systemd is not the init manager for this image") class SystemdTest(oeRuntimeTest): def systemctl(self, action = '', target = '', expected = 0, verbose = False): command = 'systemctl %s %s' % (action, target) status, output = self.target.run(command) message = '\n'.join([command, output]) if status != expected and verbose: message += self.target.run('systemctl status --full %s' % target)[1] self.assertEqual(status, expected, message) return output class SystemdBasicTests(SystemdTest): @skipUnlessPassed('test_ssh') def test_systemd_basic(self): self.systemctl('--version') @skipUnlessPassed('test_system_basic') def test_systemd_list(self): self.systemctl('list-unit-files') def settle(self): """ Block until systemd has finished activating any units being activated, or until two minutes has elapsed. Returns a tuple, either (True, '') if all units have finished activating, or (False, message string) if there are still units activating (generally, failing units that restart). """ import time endtime = time.time() + (60 * 2) while True: status, output = self.target.run('systemctl --state=activating') if "0 loaded units listed" in output: return (True, '') if time.time() >= endtime: return (False, output) time.sleep(10) @skipUnlessPassed('test_systemd_basic') def test_systemd_failed(self): settled, output = self.settle() self.assertTrue(settled, msg="Timed out waiting for systemd to settle:\n" + output) output = self.systemctl('list-units', '--failed') match = re.search("0 loaded units listed", output) if not match: output += self.systemctl('status --full --failed') self.assertTrue(match, msg="Some systemd units failed:\n%s" % output) class SystemdServiceTests(SystemdTest): @skipUnlessPassed('test_systemd_basic') def test_systemd_status(self): self.systemctl('status --full', 'avahi-daemon.service') @skipUnlessPassed('test_systemd_status') def test_systemd_stop_start(self): self.systemctl('stop', 'avahi-daemon.service') self.systemctl('is-active', 'avahi-daemon.service', expected=3, verbose=True) self.systemctl('start','avahi-daemon.service') self.systemctl('is-active', 'avahi-daemon.service', verbose=True) @skipUnlessPassed('test_systemd_basic') def test_systemd_disable_enable(self): self.systemctl('disable', 'avahi-daemon.service') self.systemctl('is-enabled', 'avahi-daemon.service', expected=1) self.systemctl('enable', 'avahi-daemon.service') self.systemctl('is-enabled', 'avahi-daemon.service') ########NEW FILE######## __FILENAME__ = vnc from oeqa.oetest import oeRuntimeTest from oeqa.utils.decorators import * import re def setUpModule(): skipModuleUnless(oeRuntimeTest.hasPackage('x11vnc'), "No x11vnc package in image") class VNCTest(oeRuntimeTest): @skipUnlessPassed('test_ssh') def test_vnc(self): (status, output) = self.target.run('x11vnc -display :0 -bg -o x11vnc.log') self.assertEqual(status, 0, msg="x11vnc server failed to start: %s" % output) port = re.search('PORT=[0-9]*', output) self.assertTrue(port, msg="Listening port not specified in command output: %s" %output) vncport = port.group(0).split('=')[1] (status, output) = self.target.run('netstat -ntl | grep ":%s"' % vncport) self.assertEqual(status, 0, msg="x11vnc server not running on port %s\n\n%s" % (vncport, self.target.run('netstat -ntl; cat x11vnc.log')[1])) ########NEW FILE######## __FILENAME__ = x32lib import unittest from oeqa.oetest import oeRuntimeTest, skipModule from oeqa.utils.decorators import * def setUpModule(): #check if DEFAULTTUNE is set and it's value is: x86-64-x32 defaulttune = oeRuntimeTest.tc.d.getVar("DEFAULTTUNE", True) if "x86-64-x32" not in defaulttune: skipModule("DEFAULTTUNE is not set to x86-64-x32") class X32libTest(oeRuntimeTest): @skipUnlessPassed("test_ssh") def test_x32_file(self): status1 = self.target.run("readelf -h /bin/ls | grep Class | grep ELF32")[0] status2 = self.target.run("readelf -h /bin/ls | grep Machine | grep X86-64")[0] self.assertTrue(status1 == 0 and status2 == 0, msg="/bin/ls isn't an X86-64 ELF32 binary. readelf says: %s" % self.target.run("readelf -h /bin/ls")[1]) ########NEW FILE######## __FILENAME__ = xorg import unittest from oeqa.oetest import oeRuntimeTest, skipModule from oeqa.utils.decorators import * def setUpModule(): if not oeRuntimeTest.hasFeature("x11-base"): skipModule("target doesn't have x11 in IMAGE_FEATURES") class XorgTest(oeRuntimeTest): @skipUnlessPassed('test_ssh') def test_xorg_running(self): (status, output) = self.target.run(oeRuntimeTest.pscmd + ' | grep -v xinit | grep [X]org') self.assertEqual(status, 0, msg="Xorg does not appear to be running %s" % self.target.run(oeRuntimeTest.pscmd)[1]) @skipUnlessPassed('test_ssh') def test_xorg_error(self): (status, output) = self.target.run('cat /var/log/Xorg.0.log | grep -v "(EE) error," | grep -v "PreInit" | grep -v "evdev:" | grep -v "glx" | grep "(EE)"') self.assertEqual(status, 1, msg="Errors in Xorg log: %s" % output) ########NEW FILE######## __FILENAME__ = base # Copyright (c) 2013 Intel Corporation # # Released under the MIT license (see COPYING.MIT) # DESCRIPTION # Base class inherited by test classes in meta/lib/selftest import unittest import os import sys import shutil import logging import errno import oeqa.utils.ftools as ftools from oeqa.utils.commands import runCmd, bitbake, get_bb_var, get_test_layer class oeSelfTest(unittest.TestCase): log = logging.getLogger("selftest.base") longMessage = True def __init__(self, methodName="runTest"): self.builddir = os.environ.get("BUILDDIR") self.localconf_path = os.path.join(self.builddir, "conf/local.conf") self.testinc_path = os.path.join(self.builddir, "conf/selftest.inc") self.testlayer_path = oeSelfTest.testlayer_path self._extra_tear_down_commands = [] self._track_for_cleanup = [] super(oeSelfTest, self).__init__(methodName) def setUp(self): os.chdir(self.builddir) # we don't know what the previous test left around in config or inc files # if it failed so we need a fresh start try: os.remove(self.testinc_path) except OSError as e: if e.errno != errno.ENOENT: raise for root, _, files in os.walk(self.testlayer_path): for f in files: if f == 'test_recipe.inc': os.remove(os.path.join(root, f)) # tests might need their own setup # but if they overwrite this one they have to call # super each time, so let's give them an alternative self.setUpLocal() def setUpLocal(self): pass def tearDown(self): if self._extra_tear_down_commands: failed_extra_commands = [] for command in self._extra_tear_down_commands: result = runCmd(command, ignore_status=True) if not result.status == 0: failed_extra_commands.append(command) if failed_extra_commands: self.log.warning("tearDown commands have failed: %s" % ', '.join(map(str, failed_extra_commands))) self.log.debug("Trying to move on.") self._extra_tear_down_commands = [] if self._track_for_cleanup: for path in self._track_for_cleanup: if os.path.isdir(path): shutil.rmtree(path) if os.path.isfile(path): os.remove(path) self._track_for_cleanup = [] self.tearDownLocal() def tearDownLocal(self): pass # add test specific commands to the tearDown method. def add_command_to_tearDown(self, command): self.log.debug("Adding command '%s' to tearDown for this test." % command) self._extra_tear_down_commands.append(command) # add test specific files or directories to be removed in the tearDown method def track_for_cleanup(self, path): self.log.debug("Adding path '%s' to be cleaned up when test is over" % path) self._track_for_cleanup.append(path) # write to <builddir>/conf/selftest.inc def write_config(self, data): self.log.debug("Writing to: %s\n%s\n" % (self.testinc_path, data)) ftools.write_file(self.testinc_path, data) # append to <builddir>/conf/selftest.inc def append_config(self, data): self.log.debug("Appending to: %s\n%s\n" % (self.testinc_path, data)) ftools.append_file(self.testinc_path, data) # remove data from <builddir>/conf/selftest.inc def remove_config(self, data): self.log.debug("Removing from: %s\n\%s\n" % (self.testinc_path, data)) ftools.remove_from_file(self.testinc_path, data) # write to meta-sefltest/recipes-test/<recipe>/test_recipe.inc def write_recipeinc(self, recipe, data): inc_file = os.path.join(self.testlayer_path, 'recipes-test', recipe, 'test_recipe.inc') self.log.debug("Writing to: %s\n%s\n" % (inc_file, data)) ftools.write_file(inc_file, data) # append data to meta-sefltest/recipes-test/<recipe>/test_recipe.inc def append_recipeinc(self, recipe, data): inc_file = os.path.join(self.testlayer_path, 'recipes-test', recipe, 'test_recipe.inc') self.log.debug("Appending to: %s\n%s\n" % (inc_file, data)) ftools.append_file(inc_file, data) # remove data from meta-sefltest/recipes-test/<recipe>/test_recipe.inc def remove_recipeinc(self, recipe, data): inc_file = os.path.join(self.testlayer_path, 'recipes-test', recipe, 'test_recipe.inc') self.log.debug("Removing from: %s\n%s\n" % (inc_file, data)) ftools.remove_from_file(inc_file, data) # delete meta-sefltest/recipes-test/<recipe>/test_recipe.inc file def delete_recipeinc(self, recipe): inc_file = os.path.join(self.testlayer_path, 'recipes-test', recipe, 'test_recipe.inc') self.log.debug("Deleting file: %s" % inc_file) try: os.remove(inc_file) except OSError as e: if e.errno != errno.ENOENT: raise ########NEW FILE######## __FILENAME__ = bblayers import unittest import os import logging import re import shutil import oeqa.utils.ftools as ftools from oeqa.selftest.base import oeSelfTest from oeqa.utils.commands import runCmd class BitbakeLayers(oeSelfTest): def test_bitbakelayers_showcrossdepends(self): result = runCmd('bitbake-layers show-cross-depends') self.assertTrue('aspell' in result.output) def test_bitbakelayers_showlayers(self): result = runCmd('bitbake-layers show_layers') self.assertTrue('meta-selftest' in result.output) def test_bitbakelayers_showappends(self): result = runCmd('bitbake-layers show_appends') self.assertTrue('xcursor-transparent-theme_0.1.1.bbappend' in result.output, msg='xcursor-transparent-theme_0.1.1.bbappend file was not recognised') def test_bitbakelayers_showoverlayed(self): result = runCmd('bitbake-layers show_overlayed') self.assertTrue('aspell' in result.output, msg='xcursor-transparent-theme_0.1.1.bbappend file was not recognised') def test_bitbakelayers_flatten(self): self.assertFalse(os.path.isdir(os.path.join(self.builddir, 'test'))) result = runCmd('bitbake-layers flatten test') bb_file = os.path.join(self.builddir, 'test/recipes-graphics/xcursor-transparent-theme/xcursor-transparent-theme_0.1.1.bb') self.assertTrue(os.path.isfile(bb_file)) contents = ftools.read_file(bb_file) find_in_contents = re.search("##### bbappended from meta-selftest #####\n(.*\n)*include test_recipe.inc", contents) shutil.rmtree(os.path.join(self.builddir, 'test')) self.assertTrue(find_in_contents) ########NEW FILE######## __FILENAME__ = bbtests import unittest import os import logging import re import shutil import oeqa.utils.ftools as ftools from oeqa.selftest.base import oeSelfTest from oeqa.utils.commands import runCmd, bitbake, get_bb_var class BitbakeTests(oeSelfTest): def test_run_bitbake_from_dir_1(self): os.chdir(os.path.join(self.builddir, 'conf')) bitbake('-e') def test_run_bitbake_from_dir_2(self): my_env = os.environ.copy() my_env['BBPATH'] = my_env['BUILDDIR'] os.chdir(os.path.dirname(os.environ['BUILDDIR'])) bitbake('-e', env=my_env) def test_event_handler(self): self.write_config("INHERIT += \"test_events\"") result = bitbake('m4-native') find_build_started = re.search("NOTE: Test for bb\.event\.BuildStarted(\n.*)*NOTE: Preparing runqueue", result.output) find_build_completed = re.search("Tasks Summary:.*(\n.*)*NOTE: Test for bb\.event\.BuildCompleted", result.output) self.assertTrue(find_build_started, msg = "Match failed in:\n%s" % result.output) self.assertTrue(find_build_completed, msg = "Match failed in:\n%s" % result.output) self.assertFalse('Test for bb.event.InvalidEvent' in result.output) def test_local_sstate(self): bitbake('m4-native -ccleansstate') bitbake('m4-native') bitbake('m4-native -cclean') result = bitbake('m4-native') find_setscene = re.search("m4-native.*do_.*_setscene", result.output) self.assertTrue(find_setscene) def test_bitbake_invalid_recipe(self): result = bitbake('-b asdf', ignore_status=True) self.assertTrue("ERROR: Unable to find any recipe file matching 'asdf'" in result.output) def test_bitbake_invalid_target(self): result = bitbake('asdf', ignore_status=True) self.assertTrue("ERROR: Nothing PROVIDES 'asdf'" in result.output) def test_warnings_errors(self): result = bitbake('-b asdf', ignore_status=True) find_warnings = re.search("Summary: There w.{2,3}? [1-9][0-9]* WARNING messages* shown", result.output) find_errors = re.search("Summary: There w.{2,3}? [1-9][0-9]* ERROR messages* shown", result.output) self.assertTrue(find_warnings, msg="Did not find the mumber of warnings at the end of the build:\n" + result.output) self.assertTrue(find_errors, msg="Did not find the mumber of errors at the end of the build:\n" + result.output) def test_invalid_patch(self): self.write_recipeinc('man', 'SRC_URI += "file://man-1.5h1-make.patch"') result = bitbake('man -c patch', ignore_status=True) self.delete_recipeinc('man') bitbake('-cclean man') self.assertTrue("ERROR: Function failed: patch_do_patch" in result.output) def test_force_task(self): bitbake('m4-native') result = bitbake('-C compile m4-native') look_for_tasks = ['do_compile', 'do_install', 'do_populate_sysroot'] for task in look_for_tasks: find_task = re.search("m4-native.*%s" % task, result.output) self.assertTrue(find_task) def test_bitbake_g(self): result = bitbake('-g core-image-full-cmdline') self.assertTrue('NOTE: PN build list saved to \'pn-buildlist\'' in result.output) self.assertTrue('openssh' in ftools.read_file(os.path.join(self.builddir, 'pn-buildlist'))) for f in ['pn-buildlist', 'pn-depends.dot', 'package-depends.dot', 'task-depends.dot']: os.remove(f) def test_image_manifest(self): bitbake('core-image-minimal') deploydir = get_bb_var("DEPLOY_DIR_IMAGE", target="core-image-minimal") imagename = get_bb_var("IMAGE_LINK_NAME", target="core-image-minimal") manifest = os.path.join(deploydir, imagename + ".manifest") self.assertTrue(os.path.islink(manifest), msg="No manifest file created for image") def test_invalid_recipe_src_uri(self): data = 'SRC_URI = "file://invalid"' self.write_recipeinc('man', data) bitbake('-ccleanall man') result = bitbake('-c fetch man', ignore_status=True) bitbake('-ccleanall man') self.delete_recipeinc('man') self.assertEqual(result.status, 1, msg='Command succeded when it should have failed') self.assertTrue('ERROR: Fetcher failure: Unable to find file file://invalid anywhere. The paths that were searched were:' in result.output) self.assertTrue('ERROR: Function failed: Fetcher failure for URL: \'file://invalid\'. Unable to fetch URL from any source.' in result.output) def test_rename_downloaded_file(self): data = 'SRC_URI_append = ";downloadfilename=test-aspell.tar.gz"' self.write_recipeinc('aspell', data) bitbake('-ccleanall aspell') result = bitbake('-c fetch aspell', ignore_status=True) self.delete_recipeinc('aspell') self.assertEqual(result.status, 0) self.assertTrue(os.path.isfile(os.path.join(get_bb_var("DL_DIR"), 'test-aspell.tar.gz'))) self.assertTrue(os.path.isfile(os.path.join(get_bb_var("DL_DIR"), 'test-aspell.tar.gz.done'))) bitbake('-ccleanall aspell') ########NEW FILE######## __FILENAME__ = buildhistory import unittest import os import re import shutil import datetime import oeqa.utils.ftools as ftools from oeqa.selftest.base import oeSelfTest from oeqa.utils.commands import Command, runCmd, bitbake, get_bb_var, get_test_layer class BuildhistoryBase(oeSelfTest): def config_buildhistory(self, tmp_bh_location=False): if (not 'buildhistory' in get_bb_var('USER_CLASSES')) and (not 'buildhistory' in get_bb_var('INHERIT')): add_buildhistory_config = 'INHERIT += "buildhistory"\nBUILDHISTORY_COMMIT = "1"' self.append_config(add_buildhistory_config) if tmp_bh_location: # Using a temporary buildhistory location for testing tmp_bh_dir = os.path.join(self.builddir, "tmp_buildhistory_%s" % datetime.datetime.now().strftime('%Y%m%d%H%M%S')) buildhistory_dir_config = "BUILDHISTORY_DIR = \"%s\"" % tmp_bh_dir self.append_config(buildhistory_dir_config) self.track_for_cleanup(tmp_bh_dir) def run_buildhistory_operation(self, target, global_config='', target_config='', change_bh_location=False, expect_error=False, error_regex=''): if change_bh_location: tmp_bh_location = True else: tmp_bh_location = False self.config_buildhistory(tmp_bh_location) self.append_config(global_config) self.append_recipeinc(target, target_config) bitbake("-cclean %s" % target) result = bitbake(target, ignore_status=True) self.remove_config(global_config) self.remove_recipeinc(target, target_config) if expect_error: self.assertEqual(result.status, 1, msg="Error expected for global config '%s' and target config '%s'" % (global_config, target_config)) search_for_error = re.search(error_regex, result.output) self.assertTrue(search_for_error, msg="Could not find desired error in output: %s" % error_regex) else: self.assertEqual(result.status, 0, msg="Command 'bitbake %s' has failed unexpectedly: %s" % (target, result.output)) ########NEW FILE######## __FILENAME__ = buildoptions import unittest import os import logging import re from oeqa.selftest.base import oeSelfTest from oeqa.selftest.buildhistory import BuildhistoryBase from oeqa.utils.commands import runCmd, bitbake, get_bb_var import oeqa.utils.ftools as ftools class ImageOptionsTests(oeSelfTest): def test_incremental_image_generation(self): bitbake("-c cleanall core-image-minimal") self.write_config('INC_RPM_IMAGE_GEN = "1"') self.append_config('IMAGE_FEATURES += "ssh-server-openssh"') bitbake("core-image-minimal") res = runCmd("grep 'Installing openssh-sshd' %s" % (os.path.join(get_bb_var("WORKDIR", "core-image-minimal"), "temp/log.do_rootfs")), ignore_status=True) self.remove_config('IMAGE_FEATURES += "ssh-server-openssh"') self.assertEqual(0, res.status, msg="No match for openssh-sshd in log.do_rootfs") bitbake("core-image-minimal") res = runCmd("grep 'Removing openssh-sshd' %s" %(os.path.join(get_bb_var("WORKDIR", "core-image-minimal"), "temp/log.do_rootfs")),ignore_status=True) self.assertEqual(0, res.status, msg="openssh-sshd was not removed from image") def test_rm_old_image(self): bitbake("core-image-minimal") deploydir = get_bb_var("DEPLOY_DIR_IMAGE", target="core-image-minimal") imagename = get_bb_var("IMAGE_LINK_NAME", target="core-image-minimal") deploydir_files = os.listdir(deploydir) track_original_files = [] for image_file in deploydir_files: if imagename in image_file and os.path.islink(os.path.join(deploydir, image_file)): track_original_files.append(os.path.realpath(os.path.join(deploydir, image_file))) self.append_config("RM_OLD_IMAGE = \"1\"") bitbake("-C rootfs core-image-minimal") deploydir_files = os.listdir(deploydir) remaining_not_expected = [path for path in track_original_files if os.path.basename(path) in deploydir_files] self.assertFalse(remaining_not_expected, msg="\nThe following image files ware not removed: %s" % ', '.join(map(str, remaining_not_expected))) def test_ccache_tool(self): bitbake("ccache-native") self.assertTrue(os.path.isfile(os.path.join(get_bb_var('STAGING_BINDIR_NATIVE', 'ccache-native'), "ccache"))) self.write_config('INHERIT += "ccache"') bitbake("m4 -c cleansstate") bitbake("m4 -c compile") res = runCmd("grep ccache %s" % (os.path.join(get_bb_var("WORKDIR","m4"),"temp/log.do_compile")), ignore_status=True) self.assertEqual(0, res.status, msg="No match for ccache in m4 log.do_compile") bitbake("ccache-native -ccleansstate") class DiskMonTest(oeSelfTest): def test_stoptask_behavior(self): self.write_config('BB_DISKMON_DIRS = "STOPTASKS,${TMPDIR},100000G,100K"') res = bitbake("m4", ignore_status = True) self.assertTrue('ERROR: No new tasks can be executed since the disk space monitor action is "STOPTASKS"!' in res.output) self.assertEqual(res.status, 1) self.write_config('BB_DISKMON_DIRS = "ABORT,${TMPDIR},100000G,100K"') res = bitbake("m4", ignore_status = True) self.assertTrue('ERROR: Immediately abort since the disk space monitor action is "ABORT"!' in res.output) self.assertEqual(res.status, 1) self.write_config('BB_DISKMON_DIRS = "WARN,${TMPDIR},100000G,100K"') res = bitbake("m4") self.assertTrue('WARNING: The free space' in res.output) class SanityOptionsTest(oeSelfTest): def test_options_warnqa_errorqa_switch(self): bitbake("xcursor-transparent-theme -ccleansstate") if "packages-list" not in get_bb_var("ERROR_QA"): self.write_config("ERROR_QA_append = \" packages-list\"") self.write_recipeinc('xcursor-transparent-theme', 'PACKAGES += \"${PN}-dbg\"') res = bitbake("xcursor-transparent-theme", ignore_status=True) self.delete_recipeinc('xcursor-transparent-theme') self.assertTrue("ERROR: QA Issue: xcursor-transparent-theme-dbg is listed in PACKAGES multiple times, this leads to packaging errors." in res.output) self.assertEqual(res.status, 1) self.write_recipeinc('xcursor-transparent-theme', 'PACKAGES += \"${PN}-dbg\"') self.append_config('ERROR_QA_remove = "packages-list"') self.append_config('WARN_QA_append = " packages-list"') res = bitbake("xcursor-transparent-theme") bitbake("xcursor-transparent-theme -ccleansstate") self.delete_recipeinc('xcursor-transparent-theme') self.assertTrue("WARNING: QA Issue: xcursor-transparent-theme-dbg is listed in PACKAGES multiple times, this leads to packaging errors." in res.output) def test_sanity_userspace_dependency(self): self.append_config('WARN_QA_append = " unsafe-references-in-binaries unsafe-references-in-scripts"') bitbake("-ccleansstate gzip nfs-utils") res = bitbake("gzip nfs-utils") self.assertTrue("WARNING: QA Issue: gzip" in res.output) self.assertTrue("WARNING: QA Issue: nfs-utils" in res.output) class BuildhistoryTests(BuildhistoryBase): def test_buildhistory_basic(self): self.run_buildhistory_operation('xcursor-transparent-theme') self.assertTrue(os.path.isdir(get_bb_var('BUILDHISTORY_DIR'))) def test_buildhistory_buildtime_pr_backwards(self): self.add_command_to_tearDown('cleanup-workdir') target = 'xcursor-transparent-theme' error = "ERROR: QA Issue: Package version for package %s went backwards which would break package feeds from (.*-r1 to .*-r0)" % target self.run_buildhistory_operation(target, target_config="PR = \"r1\"", change_bh_location=True) self.run_buildhistory_operation(target, target_config="PR = \"r0\"", change_bh_location=False, expect_error=True, error_regex=error) ########NEW FILE######## __FILENAME__ = oescripts import datetime import unittest import os import re import shutil import oeqa.utils.ftools as ftools from oeqa.selftest.base import oeSelfTest from oeqa.selftest.buildhistory import BuildhistoryBase from oeqa.utils.commands import Command, runCmd, bitbake, get_bb_var, get_test_layer class TestScripts(oeSelfTest): def test_cleanup_workdir(self): path = os.path.dirname(get_bb_var('WORKDIR', 'gzip')) old_version_recipe = os.path.join(get_bb_var('COREBASE'), 'meta/recipes-extended/gzip/gzip_1.3.12.bb') old_version = '1.3.12' bitbake("-ccleansstate gzip") bitbake("-ccleansstate -b %s" % old_version_recipe) if os.path.exists(get_bb_var('WORKDIR', "-b %s" % old_version_recipe)): shutil.rmtree(get_bb_var('WORKDIR', "-b %s" % old_version_recipe)) if os.path.exists(get_bb_var('WORKDIR', 'gzip')): shutil.rmtree(get_bb_var('WORKDIR', 'gzip')) if os.path.exists(path): initial_contents = os.listdir(path) else: initial_contents = [] bitbake('gzip') intermediary_contents = os.listdir(path) bitbake("-b %s" % old_version_recipe) runCmd('cleanup-workdir') remaining_contents = os.listdir(path) expected_contents = [x for x in intermediary_contents if x not in initial_contents] remaining_not_expected = [x for x in remaining_contents if x not in expected_contents] self.assertFalse(remaining_not_expected, msg="Not all necessary content has been deleted from %s: %s" % (path, ', '.join(map(str, remaining_not_expected)))) expected_not_remaining = [x for x in expected_contents if x not in remaining_contents] self.assertFalse(expected_not_remaining, msg="The script removed extra contents from %s: %s" % (path, ', '.join(map(str, expected_not_remaining)))) class BuildhistoryDiffTests(BuildhistoryBase): def test_buildhistory_diff(self): self.add_command_to_tearDown('cleanup-workdir') target = 'xcursor-transparent-theme' self.run_buildhistory_operation(target, target_config="PR = \"r1\"", change_bh_location=True) self.run_buildhistory_operation(target, target_config="PR = \"r0\"", change_bh_location=False, expect_error=True) result = runCmd("buildhistory-diff -p %s" % get_bb_var('BUILDHISTORY_DIR')) expected_output = 'PR changed from "r1" to "r0"' self.assertTrue(expected_output in result.output, msg="Did not find expected output: %s" % result.output) ########NEW FILE######## __FILENAME__ = prservice import unittest import os import logging import re import shutil import datetime import oeqa.utils.ftools as ftools from oeqa.selftest.base import oeSelfTest from oeqa.utils.commands import runCmd, bitbake, get_bb_var class BitbakePrTests(oeSelfTest): def get_pr_version(self, package_name): pkgdata_dir = get_bb_var('PKGDATA_DIR') package_data_file = os.path.join(pkgdata_dir, 'runtime', package_name) package_data = ftools.read_file(package_data_file) find_pr = re.search("PKGR: r[0-9]+\.([0-9]+)", package_data) self.assertTrue(find_pr) return int(find_pr.group(1)) def get_task_stamp(self, package_name, recipe_task): stampdata = get_bb_var('STAMP', target=package_name).split('/') prefix = stampdata[-1] package_stamps_path = "/".join(stampdata[:-1]) stamps = [] for stamp in os.listdir(package_stamps_path): find_stamp = re.match("%s\.%s\.([a-z0-9]{32})" % (prefix, recipe_task), stamp) if find_stamp: stamps.append(find_stamp.group(1)) self.assertFalse(len(stamps) == 0, msg="Cound not find stamp for task %s for recipe %s" % (recipe_task, package_name)) self.assertFalse(len(stamps) > 1, msg="Found multiple %s stamps for the %s recipe in the %s directory." % (recipe_task, package_name, package_stamps_path)) return str(stamps[0]) def increment_package_pr(self, package_name): inc_data = "do_package_append() {\nbb.build.exec_func('do_test_prserv', d)\n}\ndo_test_prserv() {\necho \"The current date is: %s\"\n}" % datetime.datetime.now() self.write_recipeinc(package_name, inc_data) bitbake("-ccleansstate %s" % package_name) res = bitbake(package_name, ignore_status=True) self.delete_recipeinc(package_name) self.assertEqual(res.status, 0, msg=res.output) self.assertTrue("NOTE: Started PRServer with DBfile" in res.output, msg=res.output) def config_pr_tests(self, package_name, package_type='rpm', pr_socket='localhost:0'): config_package_data = 'PACKAGE_CLASSES = "package_%s"' % package_type self.write_config(config_package_data) config_server_data = 'PRSERV_HOST = "%s"' % pr_socket self.append_config(config_server_data) def run_test_pr_service(self, package_name, package_type='rpm', track_task='do_package', pr_socket='localhost:0'): self.config_pr_tests(package_name, package_type, pr_socket) self.increment_package_pr(package_name) pr_1 = self.get_pr_version(package_name) stamp_1 = self.get_task_stamp(package_name, track_task) self.increment_package_pr(package_name) pr_2 = self.get_pr_version(package_name) stamp_2 = self.get_task_stamp(package_name, track_task) bitbake("-ccleansstate %s" % package_name) self.assertTrue(pr_2 - pr_1 == 1) self.assertTrue(stamp_1 != stamp_2) def run_test_pr_export_import(self, package_name, replace_current_db=True): self.config_pr_tests(package_name) self.increment_package_pr(package_name) pr_1 = self.get_pr_version(package_name) exported_db_path = os.path.join(self.builddir, 'export.inc') export_result = runCmd("bitbake-prserv-tool export %s" % exported_db_path, ignore_status=True) self.assertEqual(export_result.status, 0, msg="PR Service database export failed: %s" % export_result.output) if replace_current_db: current_db_path = os.path.join(get_bb_var('PERSISTENT_DIR'), 'prserv.sqlite3') self.assertTrue(os.path.exists(current_db_path), msg="Path to current PR Service database is invalid: %s" % current_db_path) os.remove(current_db_path) import_result = runCmd("bitbake-prserv-tool import %s" % exported_db_path, ignore_status=True) os.remove(exported_db_path) self.assertEqual(import_result.status, 0, msg="PR Service database import failed: %s" % import_result.output) self.increment_package_pr(package_name) pr_2 = self.get_pr_version(package_name) bitbake("-ccleansstate %s" % package_name) self.assertTrue(pr_2 - pr_1 == 1) def test_import_export_replace_db(self): self.run_test_pr_export_import('m4') def test_import_export_override_db(self): self.run_test_pr_export_import('m4', replace_current_db=False) def test_pr_service_rpm_arch_dep(self): self.run_test_pr_service('m4', 'rpm', 'do_package') def test_pr_service_deb_arch_dep(self): self.run_test_pr_service('m4', 'deb', 'do_package') def test_pr_service_ipk_arch_dep(self): self.run_test_pr_service('m4', 'ipk', 'do_package') def test_pr_service_rpm_arch_indep(self): self.run_test_pr_service('xcursor-transparent-theme', 'rpm', 'do_package') def test_pr_service_deb_arch_indep(self): self.run_test_pr_service('xcursor-transparent-theme', 'deb', 'do_package') def test_pr_service_ipk_arch_indep(self): self.run_test_pr_service('xcursor-transparent-theme', 'ipk', 'do_package') ########NEW FILE######## __FILENAME__ = sstate import datetime import unittest import os import re import shutil import oeqa.utils.ftools as ftools from oeqa.selftest.base import oeSelfTest from oeqa.utils.commands import runCmd, bitbake, get_bb_var, get_test_layer class SStateBase(oeSelfTest): def setUpLocal(self): self.temp_sstate_location = None self.sstate_path = get_bb_var('SSTATE_DIR') self.distro = get_bb_var('NATIVELSBSTRING') self.distro_specific_sstate = os.path.join(self.sstate_path, self.distro) # Creates a special sstate configuration with the option to add sstate mirrors def config_sstate(self, temp_sstate_location=False, add_local_mirrors=[]): self.temp_sstate_location = temp_sstate_location if self.temp_sstate_location: temp_sstate_path = os.path.join(self.builddir, "temp_sstate_%s" % datetime.datetime.now().strftime('%Y%m%d%H%M%S')) config_temp_sstate = "SSTATE_DIR = \"%s\"" % temp_sstate_path self.append_config(config_temp_sstate) self.track_for_cleanup(temp_sstate_path) self.sstate_path = get_bb_var('SSTATE_DIR') self.distro = get_bb_var('NATIVELSBSTRING') self.distro_specific_sstate = os.path.join(self.sstate_path, self.distro) if add_local_mirrors: config_set_sstate_if_not_set = 'SSTATE_MIRRORS ?= ""' self.append_config(config_set_sstate_if_not_set) for local_mirror in add_local_mirrors: self.assertFalse(os.path.join(local_mirror) == os.path.join(self.sstate_path), msg='Cannot add the current sstate path as a sstate mirror') config_sstate_mirror = "SSTATE_MIRRORS += \"file://.* file:///%s/PATH\"" % local_mirror self.append_config(config_sstate_mirror) # Returns a list containing sstate files def search_sstate(self, filename_regex, distro_specific=True, distro_nonspecific=True): result = [] for root, dirs, files in os.walk(self.sstate_path): if distro_specific and re.search("%s/[a-z0-9]{2}$" % self.distro, root): for f in files: if re.search(filename_regex, f): result.append(f) if distro_nonspecific and re.search("%s/[a-z0-9]{2}$" % self.sstate_path, root): for f in files: if re.search(filename_regex, f): result.append(f) return result ########NEW FILE######## __FILENAME__ = sstatetests import datetime import unittest import os import re import shutil import oeqa.utils.ftools as ftools from oeqa.selftest.base import oeSelfTest from oeqa.utils.commands import runCmd, bitbake, get_bb_var, get_test_layer from oeqa.selftest.sstate import SStateBase class SStateTests(SStateBase): # Test sstate files creation and their location def run_test_sstate_creation(self, targets, distro_specific=True, distro_nonspecific=True, temp_sstate_location=True, should_pass=True): self.config_sstate(temp_sstate_location) if self.temp_sstate_location: bitbake(['-cclean'] + targets) else: bitbake(['-ccleansstate'] + targets) bitbake(targets) file_tracker = self.search_sstate('|'.join(map(str, targets)), distro_specific, distro_nonspecific) if should_pass: self.assertTrue(file_tracker , msg="Could not find sstate files for: %s" % ', '.join(map(str, targets))) else: self.assertTrue(not file_tracker , msg="Found sstate files in the wrong place for: %s" % ', '.join(map(str, targets))) def test_sstate_creation_distro_specific_pass(self): targetarch = get_bb_var('TUNE_ARCH') self.run_test_sstate_creation(['binutils-cross-'+ targetarch, 'binutils-native'], distro_specific=True, distro_nonspecific=False, temp_sstate_location=True) def test_sstate_creation_distro_specific_fail(self): targetarch = get_bb_var('TUNE_ARCH') self.run_test_sstate_creation(['binutils-cross-'+ targetarch, 'binutils-native'], distro_specific=False, distro_nonspecific=True, temp_sstate_location=True, should_pass=False) def test_sstate_creation_distro_nonspecific_pass(self): self.run_test_sstate_creation(['eglibc-initial'], distro_specific=False, distro_nonspecific=True, temp_sstate_location=True) def test_sstate_creation_distro_nonspecific_fail(self): self.run_test_sstate_creation(['eglibc-initial'], distro_specific=True, distro_nonspecific=False, temp_sstate_location=True, should_pass=False) # Test the sstate files deletion part of the do_cleansstate task def run_test_cleansstate_task(self, targets, distro_specific=True, distro_nonspecific=True, temp_sstate_location=True): self.config_sstate(temp_sstate_location) bitbake(['-ccleansstate'] + targets) bitbake(targets) tgz_created = self.search_sstate('|'.join(map(str, [s + '.*?\.tgz$' for s in targets])), distro_specific, distro_nonspecific) self.assertTrue(tgz_created, msg="Could not find sstate .tgz files for: %s" % ', '.join(map(str, targets))) siginfo_created = self.search_sstate('|'.join(map(str, [s + '.*?\.siginfo$' for s in targets])), distro_specific, distro_nonspecific) self.assertTrue(siginfo_created, msg="Could not find sstate .siginfo files for: %s" % ', '.join(map(str, targets))) bitbake(['-ccleansstate'] + targets) tgz_removed = self.search_sstate('|'.join(map(str, [s + '.*?\.tgz$' for s in targets])), distro_specific, distro_nonspecific) self.assertTrue(not tgz_removed, msg="do_cleansstate didn't remove .tgz sstate files for: %s" % ', '.join(map(str, targets))) def test_cleansstate_task_distro_specific_nonspecific(self): targetarch = get_bb_var('TUNE_ARCH') self.run_test_cleansstate_task(['binutils-cross-' + targetarch, 'binutils-native', 'eglibc-initial'], distro_specific=True, distro_nonspecific=True, temp_sstate_location=True) def test_cleansstate_task_distro_nonspecific(self): self.run_test_cleansstate_task(['eglibc-initial'], distro_specific=False, distro_nonspecific=True, temp_sstate_location=True) def test_cleansstate_task_distro_specific(self): targetarch = get_bb_var('TUNE_ARCH') self.run_test_cleansstate_task(['binutils-cross-'+ targetarch, 'binutils-native', 'eglibc-initial'], distro_specific=True, distro_nonspecific=False, temp_sstate_location=True) # Test rebuilding of distro-specific sstate files def run_test_rebuild_distro_specific_sstate(self, targets, temp_sstate_location=True): self.config_sstate(temp_sstate_location) bitbake(['-ccleansstate'] + targets) bitbake(targets) self.assertTrue(self.search_sstate('|'.join(map(str, [s + '.*?\.tgz$' for s in targets])), distro_specific=False, distro_nonspecific=True) == [], msg="Found distro non-specific sstate for: %s" % ', '.join(map(str, targets))) file_tracker_1 = self.search_sstate('|'.join(map(str, [s + '.*?\.tgz$' for s in targets])), distro_specific=True, distro_nonspecific=False) self.assertTrue(len(file_tracker_1) >= len(targets), msg = "Not all sstate files ware created for: %s" % ', '.join(map(str, targets))) self.track_for_cleanup(self.distro_specific_sstate + "_old") shutil.copytree(self.distro_specific_sstate, self.distro_specific_sstate + "_old") shutil.rmtree(self.distro_specific_sstate) bitbake(['-cclean'] + targets) bitbake(targets) file_tracker_2 = self.search_sstate('|'.join(map(str, [s + '.*?\.tgz$' for s in targets])), distro_specific=True, distro_nonspecific=False) self.assertTrue(len(file_tracker_2) >= len(targets), msg = "Not all sstate files ware created for: %s" % ', '.join(map(str, targets))) not_recreated = [x for x in file_tracker_1 if x not in file_tracker_2] self.assertTrue(not_recreated == [], msg="The following sstate files ware not recreated: %s" % ', '.join(map(str, not_recreated))) created_once = [x for x in file_tracker_2 if x not in file_tracker_1] self.assertTrue(created_once == [], msg="The following sstate files ware created only in the second run: %s" % ', '.join(map(str, created_once))) def test_rebuild_distro_specific_sstate_cross_native_targets(self): targetarch = get_bb_var('TUNE_ARCH') self.run_test_rebuild_distro_specific_sstate(['binutils-cross-' + targetarch, 'binutils-native'], temp_sstate_location=True) def test_rebuild_distro_specific_sstate_cross_target(self): targetarch = get_bb_var('TUNE_ARCH') self.run_test_rebuild_distro_specific_sstate(['binutils-cross-' + targetarch], temp_sstate_location=True) def test_rebuild_distro_specific_sstate_native_target(self): self.run_test_rebuild_distro_specific_sstate(['binutils-native'], temp_sstate_location=True) # Test the sstate-cache-management script. Each element in the global_config list is used with the corresponding element in the target_config list # global_config elements are expected to not generate any sstate files that would be removed by sstate-cache-management.sh (such as changing the value of MACHINE) def run_test_sstate_cache_management_script(self, target, global_config=[''], target_config=[''], ignore_patterns=[]): self.assertTrue(global_config) self.assertTrue(target_config) self.assertTrue(len(global_config) == len(target_config), msg='Lists global_config and target_config should have the same number of elements') self.config_sstate(temp_sstate_location=True, add_local_mirrors=[self.sstate_path]) # If buildhistory is enabled, we need to disable version-going-backwards QA checks for this test. It may report errors otherwise. if ('buildhistory' in get_bb_var('USER_CLASSES')) or ('buildhistory' in get_bb_var('INHERIT')): remove_errors_config = 'ERROR_QA_remove = "version-going-backwards"' self.append_config(remove_errors_config) # For not this only checks if random sstate tasks are handled correctly as a group. # In the future we should add control over what tasks we check for. sstate_archs_list = [] expected_remaining_sstate = [] for idx in range(len(target_config)): self.append_config(global_config[idx]) self.append_recipeinc(target, target_config[idx]) sstate_arch = get_bb_var('SSTATE_PKGARCH', target) if not sstate_arch in sstate_archs_list: sstate_archs_list.append(sstate_arch) if target_config[idx] == target_config[-1]: target_sstate_before_build = self.search_sstate(target + '.*?\.tgz$') bitbake("-cclean %s" % target) result = bitbake(target, ignore_status=True) if target_config[idx] == target_config[-1]: target_sstate_after_build = self.search_sstate(target + '.*?\.tgz$') expected_remaining_sstate += [x for x in target_sstate_after_build if x not in target_sstate_before_build if not any(pattern in x for pattern in ignore_patterns)] self.remove_config(global_config[idx]) self.remove_recipeinc(target, target_config[idx]) self.assertEqual(result.status, 0) runCmd("sstate-cache-management.sh -y --cache-dir=%s --remove-duplicated --extra-archs=%s" % (self.sstate_path, ','.join(map(str, sstate_archs_list)))) actual_remaining_sstate = [x for x in self.search_sstate(target + '.*?\.tgz$') if not any(pattern in x for pattern in ignore_patterns)] actual_not_expected = [x for x in actual_remaining_sstate if x not in expected_remaining_sstate] self.assertFalse(actual_not_expected, msg="Files should have been removed but ware not: %s" % ', '.join(map(str, actual_not_expected))) expected_not_actual = [x for x in expected_remaining_sstate if x not in actual_remaining_sstate] self.assertFalse(expected_not_actual, msg="Extra files ware removed: %s" ', '.join(map(str, expected_not_actual))) def test_sstate_cache_management_script_using_pr_1(self): global_config = [] target_config = [] global_config.append('') target_config.append('PR = "0"') self.run_test_sstate_cache_management_script('m4', global_config, target_config, ignore_patterns=['populate_lic']) def test_sstate_cache_management_script_using_pr_2(self): global_config = [] target_config = [] global_config.append('') target_config.append('PR = "0"') global_config.append('') target_config.append('PR = "1"') self.run_test_sstate_cache_management_script('m4', global_config, target_config, ignore_patterns=['populate_lic']) def test_sstate_cache_management_script_using_pr_3(self): global_config = [] target_config = [] global_config.append('MACHINE = "qemux86-64"') target_config.append('PR = "0"') global_config.append(global_config[0]) target_config.append('PR = "1"') global_config.append('MACHINE = "qemux86"') target_config.append('PR = "1"') self.run_test_sstate_cache_management_script('m4', global_config, target_config, ignore_patterns=['populate_lic']) def test_sstate_cache_management_script_using_machine(self): global_config = [] target_config = [] global_config.append('MACHINE = "qemux86-64"') target_config.append('') global_config.append('MACHINE = "qemux86"') target_config.append('') self.run_test_sstate_cache_management_script('m4', global_config, target_config, ignore_patterns=['populate_lic']) ########NEW FILE######## __FILENAME__ = _sstatetests_noauto import datetime import unittest import os import re import shutil import oeqa.utils.ftools as ftools from oeqa.selftest.base import oeSelfTest from oeqa.utils.commands import runCmd, bitbake, get_bb_var, get_test_layer from oeqa.selftest.sstate import SStateBase class RebuildFromSState(SStateBase): @classmethod def setUpClass(self): self.builddir = os.path.join(os.environ.get('BUILDDIR')) def get_dep_targets(self, primary_targets): found_targets = [] bitbake("-g " + ' '.join(map(str, primary_targets))) with open(os.path.join(self.builddir, 'pn-buildlist'), 'r') as pnfile: found_targets = pnfile.read().splitlines() return found_targets def configure_builddir(self, builddir): os.mkdir(builddir) self.track_for_cleanup(builddir) os.mkdir(os.path.join(builddir, 'conf')) shutil.copyfile(os.path.join(os.environ.get('BUILDDIR'), 'conf/local.conf'), os.path.join(builddir, 'conf/local.conf')) config = {} config['default_sstate_dir'] = "SSTATE_DIR ?= \"${TOPDIR}/sstate-cache\"" config['null_sstate_mirrors'] = "SSTATE_MIRRORS = \"\"" config['default_tmp_dir'] = "TMPDIR = \"${TOPDIR}/tmp\"" for key in config: ftools.append_file(os.path.join(builddir, 'conf/selftest.inc'), config[key]) shutil.copyfile(os.path.join(os.environ.get('BUILDDIR'), 'conf/bblayers.conf'), os.path.join(builddir, 'conf/bblayers.conf')) try: shutil.copyfile(os.path.join(os.environ.get('BUILDDIR'), 'conf/auto.conf'), os.path.join(builddir, 'conf/auto.conf')) except: pass def hardlink_tree(self, src, dst): os.mkdir(dst) self.track_for_cleanup(dst) for root, dirs, files in os.walk(src): if root == src: continue os.mkdir(os.path.join(dst, root.split(src)[1][1:])) for sstate_file in files: os.link(os.path.join(root, sstate_file), os.path.join(dst, root.split(src)[1][1:], sstate_file)) def run_test_sstate_rebuild(self, primary_targets, relocate=False, rebuild_dependencies=False): buildA = os.path.join(self.builddir, 'buildA') if relocate: buildB = os.path.join(self.builddir, 'buildB') else: buildB = buildA if rebuild_dependencies: rebuild_targets = self.get_dep_targets(primary_targets) else: rebuild_targets = primary_targets self.configure_builddir(buildA) runCmd((". %s/oe-init-build-env %s && " % (get_bb_var('COREBASE'), buildA)) + 'bitbake ' + ' '.join(map(str, primary_targets)), shell=True, executable='/bin/bash') self.hardlink_tree(os.path.join(buildA, 'sstate-cache'), os.path.join(self.builddir, 'sstate-cache-buildA')) shutil.rmtree(buildA) failed_rebuild = [] failed_cleansstate = [] for target in rebuild_targets: self.configure_builddir(buildB) self.hardlink_tree(os.path.join(self.builddir, 'sstate-cache-buildA'), os.path.join(buildB, 'sstate-cache')) result_cleansstate = runCmd((". %s/oe-init-build-env %s && " % (get_bb_var('COREBASE'), buildB)) + 'bitbake -ccleansstate ' + target, ignore_status=True, shell=True, executable='/bin/bash') if not result_cleansstate.status == 0: failed_cleansstate.append(target) shutil.rmtree(buildB) continue result_build = runCmd((". %s/oe-init-build-env %s && " % (get_bb_var('COREBASE'), buildB)) + 'bitbake ' + target, ignore_status=True, shell=True, executable='/bin/bash') if not result_build.status == 0: failed_rebuild.append(target) shutil.rmtree(buildB) self.assertFalse(failed_rebuild, msg="The following recipes have failed to rebuild: %s" % ' '.join(map(str, failed_rebuild))) self.assertFalse(failed_cleansstate, msg="The following recipes have failed cleansstate(all others have passed both cleansstate and rebuild from sstate tests): %s" % ' '.join(map(str, failed_cleansstate))) def test_sstate_relocation(self): self.run_test_sstate_rebuild(['core-image-sato-sdk'], relocate=True, rebuild_dependencies=True) def test_sstate_rebuild(self): self.run_test_sstate_rebuild(['core-image-sato-sdk'], relocate=False, rebuild_dependencies=True) ########NEW FILE######## __FILENAME__ = targetcontrol # Copyright (C) 2013 Intel Corporation # # Released under the MIT license (see COPYING.MIT) # This module is used by testimage.bbclass for setting up and controlling a target machine. import os import shutil import subprocess import bb import traceback import sys from oeqa.utils.sshcontrol import SSHControl from oeqa.utils.qemurunner import QemuRunner from oeqa.controllers.testtargetloader import TestTargetLoader from abc import ABCMeta, abstractmethod def get_target_controller(d): testtarget = d.getVar("TEST_TARGET", True) # old, simple names if testtarget == "qemu": return QemuTarget(d) elif testtarget == "simpleremote": return SimpleRemoteTarget(d) else: # use the class name try: # is it a core class defined here? controller = getattr(sys.modules[__name__], testtarget) except AttributeError: # nope, perhaps a layer defined one try: bbpath = d.getVar("BBPATH", True).split(':') testtargetloader = TestTargetLoader() controller = testtargetloader.get_controller_module(testtarget, bbpath) except ImportError as e: bb.fatal("Failed to import {0} from available controller modules:\n{1}".format(testtarget,traceback.format_exc())) except AttributeError as e: bb.fatal("Invalid TEST_TARGET - " + str(e)) return controller(d) class BaseTarget(object): __metaclass__ = ABCMeta def __init__(self, d): self.connection = None self.ip = None self.server_ip = None self.datetime = d.getVar('DATETIME', True) self.testdir = d.getVar("TEST_LOG_DIR", True) self.pn = d.getVar("PN", True) @abstractmethod def deploy(self): self.sshlog = os.path.join(self.testdir, "ssh_target_log.%s" % self.datetime) sshloglink = os.path.join(self.testdir, "ssh_target_log") if os.path.islink(sshloglink): os.unlink(sshloglink) os.symlink(self.sshlog, sshloglink) bb.note("SSH log file: %s" % self.sshlog) @abstractmethod def start(self, params=None): pass @abstractmethod def stop(self): pass def restart(self, params=None): self.stop() self.start(params) def run(self, cmd, timeout=None): return self.connection.run(cmd, timeout) def copy_to(self, localpath, remotepath): return self.connection.copy_to(localpath, remotepath) def copy_from(self, remotepath, localpath): return self.connection.copy_from(remotepath, localpath) class QemuTarget(BaseTarget): def __init__(self, d): super(QemuTarget, self).__init__(d) self.qemulog = os.path.join(self.testdir, "qemu_boot_log.%s" % self.datetime) self.origrootfs = os.path.join(d.getVar("DEPLOY_DIR_IMAGE", True), d.getVar("IMAGE_LINK_NAME", True) + '.ext3') self.rootfs = os.path.join(self.testdir, d.getVar("IMAGE_LINK_NAME", True) + '-testimage.ext3') self.runner = QemuRunner(machine=d.getVar("MACHINE", True), rootfs=self.rootfs, tmpdir = d.getVar("TMPDIR", True), deploy_dir_image = d.getVar("DEPLOY_DIR_IMAGE", True), display = d.getVar("BB_ORIGENV", False).getVar("DISPLAY", True), logfile = self.qemulog, boottime = int(d.getVar("TEST_QEMUBOOT_TIMEOUT", True))) def deploy(self): try: shutil.copyfile(self.origrootfs, self.rootfs) except Exception as e: bb.fatal("Error copying rootfs: %s" % e) qemuloglink = os.path.join(self.testdir, "qemu_boot_log") if os.path.islink(qemuloglink): os.unlink(qemuloglink) os.symlink(self.qemulog, qemuloglink) bb.note("rootfs file: %s" % self.rootfs) bb.note("Qemu log file: %s" % self.qemulog) super(QemuTarget, self).deploy() def start(self, params=None): if self.runner.start(params): self.ip = self.runner.ip self.server_ip = self.runner.server_ip self.connection = SSHControl(ip=self.ip, logfile=self.sshlog) else: self.stop() raise bb.build.FuncFailed("%s - FAILED to start qemu - check the task log and the boot log" % self.pn) def stop(self): self.runner.stop() self.connection = None self.ip = None self.server_ip = None def restart(self, params=None): if self.runner.restart(params): self.ip = self.runner.ip self.server_ip = self.runner.server_ip self.connection = SSHControl(ip=self.ip, logfile=self.sshlog) else: raise bb.build.FuncFailed("%s - FAILED to re-start qemu - check the task log and the boot log" % self.pn) class SimpleRemoteTarget(BaseTarget): def __init__(self, d): super(SimpleRemoteTarget, self).__init__(d) addr = d.getVar("TEST_TARGET_IP", True) or bb.fatal('Please set TEST_TARGET_IP with the IP address of the machine you want to run the tests on.') self.ip = addr.split(":")[0] try: self.port = addr.split(":")[1] except IndexError: self.port = None bb.note("Target IP: %s" % self.ip) self.server_ip = d.getVar("TEST_SERVER_IP", True) if not self.server_ip: try: self.server_ip = subprocess.check_output(['ip', 'route', 'get', self.ip ]).split("\n")[0].split()[-1] except Exception as e: bb.fatal("Failed to determine the host IP address (alternatively you can set TEST_SERVER_IP with the IP address of this machine): %s" % e) bb.note("Server IP: %s" % self.server_ip) def deploy(self): super(SimpleRemoteTarget, self).deploy() def start(self, params=None): self.connection = SSHControl(self.ip, logfile=self.sshlog, port=self.port) def stop(self): self.connection = None self.ip = None self.server_ip = None ########NEW FILE######## __FILENAME__ = commands # Copyright (c) 2013-2014 Intel Corporation # # Released under the MIT license (see COPYING.MIT) # DESCRIPTION # This module is mainly used by scripts/oe-selftest and modules under meta/oeqa/selftest # It provides a class and methods for running commands on the host in a convienent way for tests. import os import sys import signal import subprocess import threading import logging from oeqa.utils import CommandError class Command(object): def __init__(self, command, bg=False, timeout=None, data=None, **options): self.defaultopts = { "stdout": subprocess.PIPE, "stderr": subprocess.STDOUT, "stdin": None, "shell": False, "bufsize": -1, } self.cmd = command self.bg = bg self.timeout = timeout self.data = data self.options = dict(self.defaultopts) if isinstance(self.cmd, basestring): self.options["shell"] = True if self.data: self.options['stdin'] = subprocess.PIPE self.options.update(options) self.status = None self.output = None self.error = None self.thread = None self.log = logging.getLogger("utils.commands") def run(self): self.process = subprocess.Popen(self.cmd, **self.options) def commThread(): self.output, self.error = self.process.communicate(self.data) self.thread = threading.Thread(target=commThread) self.thread.start() self.log.debug("Running command '%s'" % self.cmd) if not self.bg: self.thread.join(self.timeout) self.stop() def stop(self): if self.thread.isAlive(): self.process.terminate() # let's give it more time to terminate gracefully before killing it self.thread.join(5) if self.thread.isAlive(): self.process.kill() self.thread.join() self.output = self.output.rstrip() self.status = self.process.poll() self.log.debug("Command '%s' returned %d as exit code." % (self.cmd, self.status)) # logging the complete output is insane # bitbake -e output is really big # and makes the log file useless if self.status: lout = "\n".join(self.output.splitlines()[-20:]) self.log.debug("Last 20 lines:\n%s" % lout) class Result(object): pass def runCmd(command, ignore_status=False, timeout=None, assert_error=True, **options): result = Result() cmd = Command(command, timeout=timeout, **options) cmd.run() result.command = command result.status = cmd.status result.output = cmd.output result.pid = cmd.process.pid if result.status and not ignore_status: if assert_error: raise AssertionError("Command '%s' returned non-zero exit status %d:\n%s" % (command, result.status, result.output)) else: raise CommandError(result.status, command, result.output) return result def bitbake(command, ignore_status=False, timeout=None, **options): if isinstance(command, basestring): cmd = "bitbake " + command else: cmd = [ "bitbake" ] + command return runCmd(cmd, ignore_status, timeout, **options) def get_bb_env(target=None): if target: return runCmd("bitbake -e %s" % target).output else: return runCmd("bitbake -e").output def get_bb_var(var, target=None): val = None bbenv = get_bb_env(target) for line in bbenv.splitlines(): if line.startswith(var + "="): val = line.split('=')[1] val = val.replace('\"','') break return val def get_test_layer(): layers = get_bb_var("BBLAYERS").split() testlayer = None for l in layers: if "/meta-selftest" in l and os.path.isdir(l): testlayer = l break return testlayer ########NEW FILE######## __FILENAME__ = decorators # Copyright (C) 2013 Intel Corporation # # Released under the MIT license (see COPYING.MIT) # Some custom decorators that can be used by unittests # Most useful is skipUnlessPassed which can be used for # creating dependecies between two test methods. from oeqa.oetest import * class skipIfFailure(object): def __init__(self,testcase): self.testcase = testcase def __call__(self,f): def wrapped_f(*args): if self.testcase in (oeTest.testFailures or oeTest.testErrors): raise unittest.SkipTest("Testcase dependency not met: %s" % self.testcase) return f(*args) wrapped_f.__name__ = f.__name__ return wrapped_f class skipIfSkipped(object): def __init__(self,testcase): self.testcase = testcase def __call__(self,f): def wrapped_f(*args): if self.testcase in oeTest.testSkipped: raise unittest.SkipTest("Testcase dependency not met: %s" % self.testcase) return f(*args) wrapped_f.__name__ = f.__name__ return wrapped_f class skipUnlessPassed(object): def __init__(self,testcase): self.testcase = testcase def __call__(self,f): def wrapped_f(*args): if self.testcase in oeTest.testSkipped or \ self.testcase in oeTest.testFailures or \ self.testcase in oeTest.testErrors: raise unittest.SkipTest("Testcase dependency not met: %s" % self.testcase) return f(*args) wrapped_f.__name__ = f.__name__ return wrapped_f ########NEW FILE######## __FILENAME__ = ftools import os import re def write_file(path, data): wdata = data.rstrip() + "\n" with open(path, "w") as f: f.write(wdata) def append_file(path, data): wdata = data.rstrip() + "\n" with open(path, "a") as f: f.write(wdata) def read_file(path): data = None with open(path) as f: data = f.read() return data def remove_from_file(path, data): lines = read_file(path).splitlines() rmdata = data.strip().splitlines() for l in rmdata: for c in range(0, lines.count(l)): i = lines.index(l) del(lines[i]) write_file(path, "\n".join(lines)) ########NEW FILE######## __FILENAME__ = httpserver import SimpleHTTPServer import multiprocessing import os class HTTPServer(SimpleHTTPServer.BaseHTTPServer.HTTPServer): def server_start(self, root_dir): os.chdir(root_dir) self.serve_forever() class HTTPRequestHandler(SimpleHTTPServer.SimpleHTTPRequestHandler): def log_message(self, format_str, *args): pass class HTTPService(object): def __init__(self, root_dir, host=''): self.root_dir = root_dir self.host = host self.port = 0 def start(self): self.server = HTTPServer((self.host, self.port), HTTPRequestHandler) if self.port == 0: self.port = self.server.server_port self.process = multiprocessing.Process(target=self.server.server_start, args=[self.root_dir]) self.process.start() def stop(self): self.server.server_close() self.process.terminate() self.process.join() ########NEW FILE######## __FILENAME__ = qemurunner # Copyright (C) 2013 Intel Corporation # # Released under the MIT license (see COPYING.MIT) # This module provides a class for starting qemu images using runqemu. # It's used by testimage.bbclass. import subprocess import os import time import signal import re import socket import select import bb class QemuRunner: def __init__(self, machine, rootfs, display, tmpdir, deploy_dir_image, logfile, boottime): # Popen object for runqemu self.runqemu = None # pid of the qemu process that runqemu will start self.qemupid = None # target ip - from the command line self.ip = None # host ip - where qemu is running self.server_ip = None self.machine = machine self.rootfs = rootfs self.display = display self.tmpdir = tmpdir self.deploy_dir_image = deploy_dir_image self.logfile = logfile self.boottime = boottime self.runqemutime = 60 self.create_socket() def create_socket(self): self.bootlog = '' self.qemusock = None try: self.server_socket = socket.socket(socket.AF_INET, socket.SOCK_STREAM) self.server_socket.setblocking(0) self.server_socket.bind(("127.0.0.1",0)) self.server_socket.listen(2) self.serverport = self.server_socket.getsockname()[1] bb.note("Created listening socket for qemu serial console on: 127.0.0.1:%s" % self.serverport) except socket.error, msg: self.server_socket.close() bb.fatal("Failed to create listening socket: %s" %msg[1]) def log(self, msg): if self.logfile: with open(self.logfile, "a") as f: f.write("%s" % msg) def start(self, qemuparams = None): if self.display: os.environ["DISPLAY"] = self.display else: bb.error("To start qemu I need a X desktop, please set DISPLAY correctly (e.g. DISPLAY=:1)") return False if not os.path.exists(self.rootfs): bb.error("Invalid rootfs %s" % self.rootfs) return False if not os.path.exists(self.tmpdir): bb.error("Invalid TMPDIR path %s" % self.tmpdir) return False else: os.environ["OE_TMPDIR"] = self.tmpdir if not os.path.exists(self.deploy_dir_image): bb.error("Invalid DEPLOY_DIR_IMAGE path %s" % self.deploy_dir_image) return False else: os.environ["DEPLOY_DIR_IMAGE"] = self.deploy_dir_image # Set this flag so that Qemu doesn't do any grabs as SDL grabs interact # badly with screensavers. os.environ["QEMU_DONT_GRAB"] = "1" self.qemuparams = 'bootparams="console=tty1 console=ttyS0,115200n8" qemuparams="-serial tcp:127.0.0.1:%s"' % self.serverport if qemuparams: self.qemuparams = self.qemuparams[:-1] + " " + qemuparams + " " + '\"' launch_cmd = 'runqemu %s %s %s' % (self.machine, self.rootfs, self.qemuparams) self.runqemu = subprocess.Popen(launch_cmd,shell=True,stdout=subprocess.PIPE,stderr=subprocess.STDOUT,preexec_fn=os.setpgrp) bb.note("runqemu started, pid is %s" % self.runqemu.pid) bb.note("waiting at most %s seconds for qemu pid" % self.runqemutime) endtime = time.time() + self.runqemutime while not self.is_alive() and time.time() < endtime: time.sleep(1) if self.is_alive(): bb.note("qemu started - qemu procces pid is %s" % self.qemupid) cmdline = '' with open('/proc/%s/cmdline' % self.qemupid) as p: cmdline = p.read() ips = re.findall("((?:[0-9]{1,3}\.){3}[0-9]{1,3})", cmdline.split("ip=")[1]) if not ips or len(ips) != 3: bb.note("Couldn't get ip from qemu process arguments! Here is the qemu command line used: %s" % cmdline) self.stop() return False else: self.ip = ips[0] self.server_ip = ips[1] bb.note("Target IP: %s" % self.ip) bb.note("Server IP: %s" % self.server_ip) bb.note("Waiting at most %d seconds for login banner" % self.boottime ) endtime = time.time() + self.boottime socklist = [self.server_socket] reachedlogin = False stopread = False while time.time() < endtime and not stopread: sread, swrite, serror = select.select(socklist, [], [], 5) for sock in sread: if sock is self.server_socket: self.qemusock, addr = self.server_socket.accept() self.qemusock.setblocking(0) socklist.append(self.qemusock) socklist.remove(self.server_socket) bb.note("Connection from %s:%s" % addr) else: data = sock.recv(1024) if data: self.log(data) self.bootlog += data if re.search("qemu.* login:", self.bootlog): stopread = True reachedlogin = True bb.note("Reached login banner") else: socklist.remove(sock) sock.close() stopread = True if not reachedlogin: bb.note("Target didn't reached login boot in %d seconds" % self.boottime) lines = "\n".join(self.bootlog.splitlines()[-5:]) bb.note("Last 5 lines of text:\n%s" % lines) bb.note("Check full boot log: %s" % self.logfile) self.stop() return False else: bb.note("Qemu pid didn't appeared in %s seconds" % self.runqemutime) output = self.runqemu.stdout self.stop() bb.note("Output from runqemu:\n%s" % output.read()) return False return self.is_alive() def stop(self): if self.runqemu: bb.note("Sending SIGTERM to runqemu") os.killpg(self.runqemu.pid, signal.SIGTERM) endtime = time.time() + self.runqemutime while self.runqemu.poll() is None and time.time() < endtime: time.sleep(1) if self.runqemu.poll() is None: bb.note("Sending SIGKILL to runqemu") os.killpg(self.runqemu.pid, signal.SIGKILL) self.runqemu = None if self.server_socket: self.server_socket.close() self.server_socket = None self.qemupid = None self.ip = None def restart(self, qemuparams = None): bb.note("Restarting qemu process") if self.runqemu.poll() is None: self.stop() self.create_socket() if self.start(qemuparams): return True return False def is_alive(self): qemu_child = self.find_child(str(self.runqemu.pid)) if qemu_child: self.qemupid = qemu_child[0] if os.path.exists("/proc/" + str(self.qemupid)): return True return False def find_child(self,parent_pid): # # Walk the process tree from the process specified looking for a qemu-system. Return its [pid'cmd] # ps = subprocess.Popen(['ps', 'axww', '-o', 'pid,ppid,command'], stdout=subprocess.PIPE).communicate()[0] processes = ps.split('\n') nfields = len(processes[0].split()) - 1 pids = {} commands = {} for row in processes[1:]: data = row.split(None, nfields) if len(data) != 3: continue if data[1] not in pids: pids[data[1]] = [] pids[data[1]].append(data[0]) commands[data[0]] = data[2] if parent_pid not in pids: return [] parents = [] newparents = pids[parent_pid] while newparents: next = [] for p in newparents: if p in pids: for n in pids[p]: if n not in parents and n not in next: next.append(n) if p not in parents: parents.append(p) newparents = next #print "Children matching %s:" % str(parents) for p in parents: # Need to be careful here since runqemu-internal runs "ldd qemu-system-xxxx" # Also, old versions of ldd (2.11) run "LD_XXXX qemu-system-xxxx" basecmd = commands[p].split()[0] basecmd = os.path.basename(basecmd) if "qemu-system" in basecmd and "-serial tcp" in commands[p]: return [int(p),commands[p]] ########NEW FILE######## __FILENAME__ = sshcontrol # Copyright (C) 2013 Intel Corporation # # Released under the MIT license (see COPYING.MIT) # Provides a class for setting up ssh connections, # running commands and copying files to/from a target. # It's used by testimage.bbclass and tests in lib/oeqa/runtime. import subprocess import time import os import select class SSHProcess(object): def __init__(self, **options): self.defaultopts = { "stdout": subprocess.PIPE, "stderr": subprocess.STDOUT, "stdin": None, "shell": False, "bufsize": -1, "preexec_fn": os.setsid, } self.options = dict(self.defaultopts) self.options.update(options) self.status = None self.output = None self.process = None self.starttime = None self.logfile = None def log(self, msg): if self.logfile: with open(self.logfile, "a") as f: f.write("%s" % msg) def run(self, command, timeout=None, logfile=None): self.logfile = logfile self.starttime = time.time() output = '' self.process = subprocess.Popen(command, **self.options) if timeout: endtime = self.starttime + timeout eof = False while time.time() < endtime and not eof: if select.select([self.process.stdout], [], [], 5)[0] != []: data = os.read(self.process.stdout.fileno(), 1024) if not data: self.process.stdout.close() eof = True else: output += data self.log(data) endtime = time.time() + timeout # process hasn't returned yet if not eof: self.process.terminate() time.sleep(5) try: self.process.kill() except OSError: pass lastline = "\nProcess killed - no output for %d seconds. Total running time: %d seconds." % (timeout, time.time() - self.starttime) self.log(lastline) output += lastline else: output = self.process.communicate()[0] self.log(output.rstrip()) self.status = self.process.wait() self.output = output.rstrip() return (self.status, self.output) class SSHControl(object): def __init__(self, ip, logfile=None, timeout=300, user='root', port=None): self.ip = ip self.defaulttimeout = timeout self.ignore_status = True self.logfile = logfile self.user = user self.ssh_options = [ '-o', 'UserKnownHostsFile=/dev/null', '-o', 'StrictHostKeyChecking=no', '-o', 'LogLevel=ERROR' ] self.ssh = ['ssh', '-l', self.user ] + self.ssh_options self.scp = ['scp'] + self.ssh_options if port: self.ssh = self.ssh + [ '-p', port ] self.scp = self.scp + [ '-P', port ] def log(self, msg): if self.logfile: with open(self.logfile, "a") as f: f.write("%s\n" % msg) def _internal_run(self, command, timeout=None, ignore_status = True): self.log("[Running]$ %s" % " ".join(command)) proc = SSHProcess() status, output = proc.run(command, timeout, logfile=self.logfile) self.log("[Command returned '%d' after %.2f seconds]" % (status, time.time() - proc.starttime)) if status and not ignore_status: raise AssertionError("Command '%s' returned non-zero exit status %d:\n%s" % (command, status, output)) return (status, output) def run(self, command, timeout=None): """ command - ssh command to run timeout=<val> - kill command if there is no output after <val> seconds timeout=None - kill command if there is no output after a default value seconds timeout=0 - no timeout, let command run until it returns """ # We need to source /etc/profile for a proper PATH on the target command = self.ssh + [self.ip, ' . /etc/profile; ' + command] if timeout is None: return self._internal_run(command, self.defaulttimeout, self.ignore_status) if timeout == 0: return self._internal_run(command, None, self.ignore_status) return self._internal_run(command, timeout, self.ignore_status) def copy_to(self, localpath, remotepath): command = self.scp + [localpath, '%s@%s:%s' % (self.user, self.ip, remotepath)] return self._internal_run(command, ignore_status=False) def copy_from(self, remotepath, localpath): command = self.scp + ['%s@%s:%s' % (self.user, self.ip, remotepath), localpath] return self._internal_run(command, ignore_status=False) ########NEW FILE######## __FILENAME__ = targetbuild # Copyright (C) 2013 Intel Corporation # # Released under the MIT license (see COPYING.MIT) # Provides a class for automating build tests for projects import os import re import subprocess class TargetBuildProject(): def __init__(self, target, d, uri, foldername=None): self.target = target self.d = d self.uri = uri self.targetdir = "~/" self.archive = os.path.basename(uri) self.localarchive = "/tmp/" + self.archive self.fname = re.sub(r'.tar.bz2|tar.gz$', '', self.archive) if foldername: self.fname = foldername def download_archive(self): exportvars = ['HTTP_PROXY', 'http_proxy', 'HTTPS_PROXY', 'https_proxy', 'FTP_PROXY', 'ftp_proxy', 'FTPS_PROXY', 'ftps_proxy', 'NO_PROXY', 'no_proxy', 'ALL_PROXY', 'all_proxy', 'SOCKS5_USER', 'SOCKS5_PASSWD'] cmd = '' for var in exportvars: val = self.d.getVar(var, True) if val: cmd = 'export ' + var + '=\"%s\"; %s' % (val, cmd) cmd = cmd + "wget -O %s %s" % (self.localarchive, self.uri) subprocess.check_call(cmd, shell=True) (status, output) = self.target.copy_to(self.localarchive, self.targetdir) if status != 0: raise Exception("Failed to copy archive to target, output: %s" % output) (status, output) = self.target.run('tar xf %s%s -C %s' % (self.targetdir, self.archive, self.targetdir)) if status != 0: raise Exception("Failed to extract archive, output: %s" % output) #Change targetdir to project folder self.targetdir = self.targetdir + self.fname # The timeout parameter of target.run is set to 0 to make the ssh command # run with no timeout. def run_configure(self): return self.target.run('cd %s; ./configure' % self.targetdir, 0)[0] def run_make(self): return self.target.run('cd %s; make' % self.targetdir, 0)[0] def run_install(self): return self.target.run('cd %s; make install' % self.targetdir, 0)[0] def clean(self): self.target.run('rm -rf %s' % self.targetdir) subprocess.call('rm -f %s' % self.localarchive, shell=True) ########NEW FILE######## __FILENAME__ = sitecustomize # OpenEmbedded sitecustomize.py (C) 2002-2008 Michael 'Mickey' Lauer <mlauer@vanille-media.de> # GPLv2 or later # Version: 20081123 # Features: # * set proper default encoding # * enable readline completion in the interactive interpreter # * load command line history on startup # * save command line history on exit import os def __exithandler(): try: readline.write_history_file( "%s/.python-history" % os.getenv( "HOME", "/tmp" ) ) except IOError: pass def __registerExitHandler(): import atexit atexit.register( __exithandler ) def __enableReadlineSupport(): readline.set_history_length( 1000 ) readline.parse_and_bind( "tab: complete" ) try: readline.read_history_file( "%s/.python-history" % os.getenv( "HOME", "/tmp" ) ) except IOError: pass def __enableDefaultEncoding(): import sys try: sys.setdefaultencoding( "utf8" ) except LookupError: pass import sys try: import rlcompleter, readline except ImportError: pass else: __enableDefaultEncoding() __registerExitHandler() __enableReadlineSupport() ########NEW FILE######## __FILENAME__ = sitecustomize # OpenEmbedded sitecustomize.py (C) 2002-2008 Michael 'Mickey' Lauer <mlauer@vanille-media.de> # GPLv2 or later # Version: 20081123 # Features: # * set proper default encoding # * enable readline completion in the interactive interpreter # * load command line history on startup # * save command line history on exit import os def __exithandler(): try: readline.write_history_file( "%s/.python-history" % os.getenv( "HOME", "/tmp" ) ) except IOError: pass def __registerExitHandler(): import atexit atexit.register( __exithandler ) def __enableReadlineSupport(): readline.set_history_length( 1000 ) readline.parse_and_bind( "tab: complete" ) try: readline.read_history_file( "%s/.python-history" % os.getenv( "HOME", "/tmp" ) ) except IOError: pass import sys try: import rlcompleter, readline except ImportError: pass else: __registerExitHandler() __enableReadlineSupport() ########NEW FILE######## __FILENAME__ = please_wait_dialog #!/usr/bin/env python import glib import gtk def destroy_window_cb(widget, event): gtk.main_quit() def all_done_cb(): gtk.main_quit() def dialogue_ui(): window = gtk.Window() window.set_title("Please wait...") window.connect("delete-event", destroy_window_cb) window.show() window.set_border_width(12) msg = "Please wait while BitBake initializes Hob" label = gtk.Label(msg) label.show() window.add(label) glib.timeout_add_seconds(10, all_done_cb) if __name__ == "__main__": dialogue_ui() gtk.main() ########NEW FILE######## __FILENAME__ = rpm-createsolvedb #!/usr/bin/env python # # This script generates a solution database for a directory containing rpm packages # but tries to be efficient about this, only doing so when the packages have changed # in some way. # # It is assumed something already went through and removed all the solvedb.done stamp files # in advance. # # First argument - the rpm binary to use # Subsequent arguments - paths to process solution databases for # import sys, os import hashlib import stat import subprocess if len(sys.argv) < 1: print("Error, rpm command not specified") sys.exit(1) if len(sys.argv) < 2: print("Error, no paths specified") sys.exit(1) paths = sys.argv[2:] for path in paths: if os.path.exists(path + "/solvedb.done"): continue data = "" manifest = [] for root, dirs, files in os.walk(path): for file in files: f = os.path.join(root, file) if f.startswith(path + "/" + "solvedb"): continue data = data + str(os.stat(f)[stat.ST_MTIME]) manifest.append(f) checksum = hashlib.md5(data).hexdigest() if os.path.exists(path + "/solvedb.checksum") and open(path + "/solvedb.checksum", "r").read() == checksum: open(path + "/solvedb.done", "w") continue if os.path.exists(path + "/solvedb"): subprocess.call("rm -rf %s" % (path + "/solvedb"), shell=True) os.mkdir(path + "/solvedb") m = open(path + "/solvedb/manifest", "w") m.write("# Dynamically generated solve manifest\n") for f in manifest: m.write(f + "\n") m.close() cmd = sys.argv[1] + ' -i --replacepkgs --replacefiles --oldpackage -D "_dbpath ' + path + '/solvedb" --justdb \ --noaid --nodeps --noorder --noscripts --notriggers --noparentdirs --nolinktos --stats \ --ignoresize --nosignature --nodigest -D "__dbi_txn create nofsync" \ ' + path + '/solvedb/manifest' subprocess.call(cmd, shell=True) open(path + "/solvedb.checksum", "w").write(checksum) open(path + "/solvedb.done", "w") ########NEW FILE######## __FILENAME__ = bbvars #!/usr/bin/env python # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation; either version 2 of the License, or # (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program; if not, write to the Free Software # Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. # # Copyright (C) Darren Hart <dvhart@linux.intel.com>, 2010 import sys import getopt import os import os.path import re def usage(): print 'Usage: %s -d FILENAME [-d FILENAME]* -m METADIR [-m MATADIR]*' % os.path.basename(sys.argv[0]) print ' -d FILENAME documentation file to search' print ' -h, --help display this help and exit' print ' -m METADIR meta directory to search for recipes' print ' -t FILENAME documentation config file (for doc tags)' print ' -T Only display variables with doc tags (requires -t)' def recipe_bbvars(recipe): ''' Return a unique set of every bbvar encountered in the recipe ''' prog = re.compile("[A-Z_]+") vset = set() try: r = open(recipe) except IOError as (errno, strerror): print 'WARNING: Failed to open recipe ', recipe print strerror for line in r: # Strip any comments from the line line = line.rsplit('#')[0] vset = vset.union(set(prog.findall(line))) r.close() bbvars = {} for v in vset: bbvars[v] = 1 return bbvars def collect_bbvars(metadir): ''' Walk the metadir and collect the bbvars from each recipe found ''' bbvars = {} for root,dirs,files in os.walk(metadir): for name in files: if name.find(".bb") >= 0: for key in recipe_bbvars(os.path.join(root,name)).iterkeys(): if bbvars.has_key(key): bbvars[key] = bbvars[key] + 1 else: bbvars[key] = 1 return bbvars def bbvar_is_documented(var, docfiles): prog = re.compile(".*($|[^A-Z_])%s([^A-Z_]|$)" % (var)) for doc in docfiles: try: f = open(doc) except IOError as (errno, strerror): print 'WARNING: Failed to open doc ', doc print strerror for line in f: if prog.match(line): return True f.close() return False def bbvar_doctag(var, docconf): prog = re.compile('^%s\[doc\] *= *"(.*)"' % (var)) if docconf == "": return "?" try: f = open(docconf) except IOError as (errno, strerror): return strerror for line in f: m = prog.search(line) if m: return m.group(1) f.close() return "" def main(): docfiles = [] metadirs = [] bbvars = {} undocumented = [] docconf = "" onlydoctags = False # Collect and validate input try: opts, args = getopt.getopt(sys.argv[1:], "d:hm:t:T", ["help"]) except getopt.GetoptError, err: print '%s' % str(err) usage() sys.exit(2) for o, a in opts: if o in ('-h', '--help'): usage() sys.exit(0) elif o == '-d': if os.path.isfile(a): docfiles.append(a) else: print 'ERROR: documentation file %s is not a regular file' % (a) sys.exit(3) elif o == '-m': if os.path.isdir(a): metadirs.append(a) else: print 'ERROR: meta directory %s is not a directory' % (a) sys.exit(4) elif o == "-t": if os.path.isfile(a): docconf = a elif o == "-T": onlydoctags = True else: assert False, "unhandled option" if len(docfiles) == 0: print 'ERROR: no docfile specified' usage() sys.exit(5) if len(metadirs) == 0: print 'ERROR: no metadir specified' usage() sys.exit(6) if onlydoctags and docconf == "": print 'ERROR: no docconf specified' usage() sys.exit(7) # Collect all the variable names from the recipes in the metadirs for m in metadirs: for key,cnt in collect_bbvars(m).iteritems(): if bbvars.has_key(key): bbvars[key] = bbvars[key] + cnt else: bbvars[key] = cnt # Check each var for documentation varlen = 0 for v in bbvars.iterkeys(): if len(v) > varlen: varlen = len(v) if not bbvar_is_documented(v, docfiles): undocumented.append(v) undocumented.sort() varlen = varlen + 1 # Report all undocumented variables print 'Found %d undocumented bb variables (out of %d):' % (len(undocumented), len(bbvars)) header = '%s%s%s' % (str("VARIABLE").ljust(varlen), str("COUNT").ljust(6), str("DOCTAG").ljust(7)) print header print str("").ljust(len(header), '=') for v in undocumented: doctag = bbvar_doctag(v, docconf) if not onlydoctags or not doctag == "": print '%s%s%s' % (v.ljust(varlen), str(bbvars[v]).ljust(6), doctag) if __name__ == "__main__": main() ########NEW FILE######## __FILENAME__ = list-packageconfig-flags #!/usr/bin/env python # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation; either version 2 of the License, or # (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program; if not, write to the Free Software Foundation. # # Copyright (C) 2013 Wind River Systems, Inc. # # - list available pkgs which have PACKAGECONFIG flags # - list available PACKAGECONFIG flags and all affected pkgs # - list all pkgs and PACKAGECONFIG information import sys import getopt import os def search_bitbakepath(): bitbakepath = "" # Search path to bitbake lib dir in order to load bb modules if os.path.exists(os.path.join(os.path.dirname(sys.argv[0]), '../../bitbake/lib/bb')): bitbakepath = os.path.join(os.path.dirname(sys.argv[0]), '../../bitbake/lib') bitbakepath = os.path.abspath(bitbakepath) else: # Look for bitbake/bin dir in PATH for pth in os.environ['PATH'].split(':'): if os.path.exists(os.path.join(pth, '../lib/bb')): bitbakepath = os.path.abspath(os.path.join(pth, '../lib')) break if not bitbakepath: sys.stderr.write("Unable to find bitbake by searching parent directory of this script or PATH\n") sys.exit(1) return bitbakepath # For importing the following modules sys.path.insert(0, search_bitbakepath()) import bb.cache import bb.cooker import bb.providers import bb.tinfoil usage_body = ''' list available pkgs which have PACKAGECONFIG flags OPTION: -h, --help display this help and exit -f, --flag list available PACKAGECONFIG flags and all affected pkgs -a, --all list all pkgs and PACKAGECONFIG information -p, --prefer list pkgs with preferred version EXAMPLE: list-packageconfig-flags.py list-packageconfig-flags.py -f list-packageconfig-flags.py -a list-packageconfig-flags.py -p list-packageconfig-flags.py -f -p list-packageconfig-flags.py -a -p ''' def usage(): print 'Usage: %s [-f|-a] [-p]' % os.path.basename(sys.argv[0]) print usage_body def get_fnlist(bbhandler, pkg_pn, preferred): ''' Get all recipe file names ''' if preferred: (latest_versions, preferred_versions) = bb.providers.findProviders(bbhandler.config_data, bbhandler.cooker.recipecache, pkg_pn) fn_list = [] for pn in sorted(pkg_pn): if preferred: fn_list.append(preferred_versions[pn][1]) else: fn_list.extend(pkg_pn[pn]) return fn_list def get_recipesdata(bbhandler, preferred): ''' Get data of all available recipes which have PACKAGECONFIG flags ''' pkg_pn = bbhandler.cooker.recipecache.pkg_pn data_dict = {} for fn in get_fnlist(bbhandler, pkg_pn, preferred): data = bb.cache.Cache.loadDataFull(fn, bbhandler.cooker.collection.get_file_appends(fn), bbhandler.config_data) if data.getVarFlags("PACKAGECONFIG"): data_dict[fn] = data return data_dict def collect_pkgs(data_dict): ''' Collect available pkgs in which have PACKAGECONFIG flags ''' # pkg_dict = {'pkg1': ['flag1', 'flag2',...]} pkg_dict = {} for fn in data_dict: pkgconfigflags = data_dict[fn].getVarFlags("PACKAGECONFIG") pkgname = data_dict[fn].getVar("P", True) pkg_dict[pkgname] = sorted(pkgconfigflags.keys()) return pkg_dict def collect_flags(pkg_dict): ''' Collect available PACKAGECONFIG flags and all affected pkgs ''' # flag_dict = {'flag': ['pkg1', 'pkg2',...]} flag_dict = {} for pkgname, flaglist in pkg_dict.iteritems(): for flag in flaglist: if flag == "defaultval": continue if flag in flag_dict: flag_dict[flag].append(pkgname) else: flag_dict[flag] = [pkgname] return flag_dict def display_pkgs(pkg_dict): ''' Display available pkgs which have PACKAGECONFIG flags ''' pkgname_len = len("PACKAGE NAME") + 1 for pkgname in pkg_dict: if pkgname_len < len(pkgname): pkgname_len = len(pkgname) pkgname_len += 1 header = '%-*s%s' % (pkgname_len, str("PACKAGE NAME"), str("PACKAGECONFIG FLAGS")) print header print str("").ljust(len(header), '=') for pkgname in sorted(pkg_dict): print('%-*s%s' % (pkgname_len, pkgname, ' '.join(pkg_dict[pkgname]))) def display_flags(flag_dict): ''' Display available PACKAGECONFIG flags and all affected pkgs ''' flag_len = len("PACKAGECONFIG FLAG") + 5 header = '%-*s%s' % (flag_len, str("PACKAGECONFIG FLAG"), str("PACKAGE NAMES")) print header print str("").ljust(len(header), '=') for flag in sorted(flag_dict): print('%-*s%s' % (flag_len, flag, ' '.join(sorted(flag_dict[flag])))) def display_all(data_dict): ''' Display all pkgs and PACKAGECONFIG information ''' print str("").ljust(50, '=') for fn in data_dict: print('%s' % data_dict[fn].getVar("P", True)) print fn packageconfig = data_dict[fn].getVar("PACKAGECONFIG", True) or '' if packageconfig.strip() == '': packageconfig = 'None' print('PACKAGECONFIG %s' % packageconfig) for flag,flag_val in data_dict[fn].getVarFlags("PACKAGECONFIG").iteritems(): if flag == "defaultval": continue print('PACKAGECONFIG[%s] %s' % (flag, flag_val)) print '' def main(): listtype = 'pkgs' preferred = False pkg_dict = {} flag_dict = {} # Collect and validate input try: opts, args = getopt.getopt(sys.argv[1:], "hfap", ["help", "flag", "all", "prefer"]) except getopt.GetoptError, err: print >> sys.stderr,'%s' % str(err) usage() sys.exit(2) for opt, value in opts: if opt in ('-h', '--help'): usage() sys.exit(0) elif opt in ('-f', '--flag'): listtype = 'flags' elif opt in ('-a', '--all'): listtype = 'all' elif opt in ('-p', '--prefer'): preferred = True else: assert False, "unhandled option" bbhandler = bb.tinfoil.Tinfoil() bbhandler.prepare() data_dict = get_recipesdata(bbhandler, preferred) if listtype == 'flags': pkg_dict = collect_pkgs(data_dict) flag_dict = collect_flags(pkg_dict) display_flags(flag_dict) elif listtype == 'pkgs': pkg_dict = collect_pkgs(data_dict) display_pkgs(pkg_dict) elif listtype == 'all': display_all(data_dict) if __name__ == "__main__": main() ########NEW FILE######## __FILENAME__ = generate-manifest-2.7 #!/usr/bin/env python # generate Python Manifest for the OpenEmbedded build system # (C) 2002-2010 Michael 'Mickey' Lauer <mlauer@vanille-media.de> # (C) 2007 Jeremy Laine # licensed under MIT, see COPYING.MIT # # June 22, 2011 -- Mark Hatle <mark.hatle@windriver.com> # * Updated to no longer generate special -dbg package, instead use the # single system -dbg # * Update version with ".1" to indicate this change import os import sys import time VERSION = "2.7.2" __author__ = "Michael 'Mickey' Lauer <mlauer@vanille-media.de>" __version__ = "20110222.2" class MakefileMaker: def __init__( self, outfile ): """initialize""" self.packages = {} self.targetPrefix = "${libdir}/python%s/" % VERSION[:3] self.output = outfile self.out( """ # WARNING: This file is AUTO GENERATED: Manual edits will be lost next time I regenerate the file. # Generator: '%s' Version %s (C) 2002-2010 Michael 'Mickey' Lauer <mlauer@vanille-media.de> # Visit the Python for Embedded Systems Site => http://www.Vanille.de/projects/python.spy """ % ( sys.argv[0], __version__ ) ) # # helper functions # def out( self, data ): """print a line to the output file""" self.output.write( "%s\n" % data ) def setPrefix( self, targetPrefix ): """set a file prefix for addPackage files""" self.targetPrefix = targetPrefix def doProlog( self ): self.out( """ """ ) self.out( "" ) def addPackage( self, name, description, dependencies, filenames ): """add a package to the Makefile""" if type( filenames ) == type( "" ): filenames = filenames.split() fullFilenames = [] for filename in filenames: if filename[0] != "$": fullFilenames.append( "%s%s" % ( self.targetPrefix, filename ) ) else: fullFilenames.append( filename ) self.packages[name] = description, dependencies, fullFilenames def doBody( self ): """generate body of Makefile""" global VERSION # # generate provides line # provideLine = 'PROVIDES+="' for name in sorted(self.packages): provideLine += "%s " % name provideLine += '"' self.out( provideLine ) self.out( "" ) # # generate package line # packageLine = 'PACKAGES="${PN}-dbg ' for name in sorted(self.packages): if name.startswith("${PN}-distutils"): if name == "${PN}-distutils": packageLine += "%s-staticdev %s " % (name, name) elif name != '${PN}-dbg': packageLine += "%s " % name packageLine += '${PN}-modules"' self.out( packageLine ) self.out( "" ) # # generate package variables # for name, data in sorted(self.packages.iteritems()): desc, deps, files = data # # write out the description, revision and dependencies # self.out( 'SUMMARY_%s="%s"' % ( name, desc ) ) self.out( 'RDEPENDS_%s="%s"' % ( name, deps ) ) line = 'FILES_%s="' % name # # check which directories to make in the temporary directory # dirset = {} # if python had a set-datatype this would be sufficient. for now, we're using a dict instead. for target in files: dirset[os.path.dirname( target )] = True # # generate which files to copy for the target (-dfR because whole directories are also allowed) # for target in files: line += "%s " % target line += '"' self.out( line ) self.out( "" ) self.out( 'SUMMARY_${PN}-modules="All Python modules"' ) line = 'RDEPENDS_${PN}-modules="' for name, data in sorted(self.packages.iteritems()): if name not in ['${PN}-dev', '${PN}-distutils-staticdev']: line += "%s " % name self.out( "%s \"" % line ) self.out( 'ALLOW_EMPTY_${PN}-modules = "1"' ) def doEpilog( self ): self.out( """""" ) self.out( "" ) def make( self ): self.doProlog() self.doBody() self.doEpilog() if __name__ == "__main__": if len( sys.argv ) > 1: try: os.unlink(sys.argv[1]) except Exception: sys.exc_clear() outfile = file( sys.argv[1], "w" ) else: outfile = sys.stdout m = MakefileMaker( outfile ) # Add packages here. Only specify dlopen-style library dependencies here, no ldd-style dependencies! # Parameters: revision, name, description, dependencies, filenames # m.addPackage( "${PN}-core", "Python interpreter and core modules", "${PN}-lang ${PN}-re", "__future__.* _abcoll.* abc.* copy.* copy_reg.* ConfigParser.* " + "genericpath.* getopt.* linecache.* new.* " + "os.* posixpath.* struct.* " + "warnings.* site.* stat.* " + "UserDict.* UserList.* UserString.* " + "lib-dynload/binascii.so lib-dynload/_struct.so lib-dynload/time.so " + "lib-dynload/xreadlines.so types.* platform.* ${bindir}/python* " + "_weakrefset.* sysconfig.* config/Makefile " + "${includedir}/python${PYTHON_MAJMIN}/pyconfig*.h " + "${libdir}/python${PYTHON_MAJMIN}/sitecustomize.py ") m.addPackage( "${PN}-dev", "Python development package", "${PN}-core", "${includedir} " + "${libdir}/lib*${SOLIBSDEV} " + "${libdir}/*.la " + "${libdir}/*.a " + "${libdir}/*.o " + "${libdir}/pkgconfig " + "${base_libdir}/*.a " + "${base_libdir}/*.o " + "${datadir}/aclocal " + "${datadir}/pkgconfig " ) m.addPackage( "${PN}-2to3", "Python automated Python 2 to 3 code translator", "${PN}-core", "${bindir}/2to3 lib2to3" ) # package m.addPackage( "${PN}-idle", "Python Integrated Development Environment", "${PN}-core ${PN}-tkinter", "${bindir}/idle idlelib" ) # package m.addPackage( "${PN}-pydoc", "Python interactive help support", "${PN}-core ${PN}-lang ${PN}-stringold ${PN}-re", "${bindir}/pydoc pydoc.* pydoc_data" ) m.addPackage( "${PN}-smtpd", "Python Simple Mail Transport Daemon", "${PN}-core ${PN}-netserver ${PN}-email ${PN}-mime", "${bindir}/smtpd.* smtpd.*" ) m.addPackage( "${PN}-audio", "Python Audio Handling", "${PN}-core", "wave.* chunk.* sndhdr.* lib-dynload/ossaudiodev.so lib-dynload/audioop.so audiodev.* sunaudio.* sunau.* toaiff.*" ) m.addPackage( "${PN}-bsddb", "Python bindings for the Berkeley Database", "${PN}-core", "bsddb lib-dynload/_bsddb.so" ) # package m.addPackage( "${PN}-codecs", "Python codecs, encodings & i18n support", "${PN}-core ${PN}-lang", "codecs.* encodings gettext.* locale.* lib-dynload/_locale.so lib-dynload/_codecs* lib-dynload/_multibytecodec.so lib-dynload/unicodedata.so stringprep.* xdrlib.*" ) m.addPackage( "${PN}-compile", "Python bytecode compilation support", "${PN}-core", "py_compile.* compileall.*" ) m.addPackage( "${PN}-compiler", "Python compiler support", "${PN}-core", "compiler" ) # package m.addPackage( "${PN}-compression", "Python high-level compression support", "${PN}-core ${PN}-zlib", "gzip.* zipfile.* tarfile.* lib-dynload/bz2.so" ) m.addPackage( "${PN}-crypt", "Python basic cryptographic and hashing support", "${PN}-core", "hashlib.* md5.* sha.* lib-dynload/crypt.so lib-dynload/_hashlib.so lib-dynload/_sha256.so lib-dynload/_sha512.so" ) m.addPackage( "${PN}-textutils", "Python option parsing, text wrapping and CSV support", "${PN}-core ${PN}-io ${PN}-re ${PN}-stringold", "lib-dynload/_csv.so csv.* optparse.* textwrap.*" ) m.addPackage( "${PN}-curses", "Python curses support", "${PN}-core", "curses lib-dynload/_curses.so lib-dynload/_curses_panel.so" ) # directory + low level module m.addPackage( "${PN}-ctypes", "Python C types support", "${PN}-core", "ctypes lib-dynload/_ctypes.so lib-dynload/_ctypes_test.so" ) # directory + low level module m.addPackage( "${PN}-datetime", "Python calendar and time support", "${PN}-core ${PN}-codecs", "_strptime.* calendar.* lib-dynload/datetime.so" ) m.addPackage( "${PN}-db", "Python file-based database support", "${PN}-core", "anydbm.* dumbdbm.* whichdb.* " ) m.addPackage( "${PN}-debugger", "Python debugger", "${PN}-core ${PN}-io ${PN}-lang ${PN}-re ${PN}-stringold ${PN}-shell ${PN}-pprint", "bdb.* pdb.*" ) m.addPackage( "${PN}-difflib", "Python helpers for computing deltas between objects", "${PN}-lang ${PN}-re", "difflib.*" ) m.addPackage( "${PN}-distutils-staticdev", "Python distribution utilities (static libraries)", "${PN}-distutils", "config/lib*.a" ) # package m.addPackage( "${PN}-distutils", "Python Distribution Utilities", "${PN}-core", "config distutils" ) # package m.addPackage( "${PN}-doctest", "Python framework for running examples in docstrings", "${PN}-core ${PN}-lang ${PN}-io ${PN}-re ${PN}-unittest ${PN}-debugger ${PN}-difflib", "doctest.*" ) # FIXME consider adding to some higher level package m.addPackage( "${PN}-elementtree", "Python elementree", "${PN}-core", "lib-dynload/_elementtree.so" ) m.addPackage( "${PN}-email", "Python email support", "${PN}-core ${PN}-io ${PN}-re ${PN}-mime ${PN}-audio ${PN}-image ${PN}-netclient", "imaplib.* email" ) # package m.addPackage( "${PN}-fcntl", "Python's fcntl interface", "${PN}-core", "lib-dynload/fcntl.so" ) m.addPackage( "${PN}-hotshot", "Python hotshot performance profiler", "${PN}-core", "hotshot lib-dynload/_hotshot.so" ) m.addPackage( "${PN}-html", "Python HTML processing support", "${PN}-core", "formatter.* htmlentitydefs.* htmllib.* markupbase.* sgmllib.* HTMLParser.* " ) m.addPackage( "${PN}-gdbm", "Python GNU database support", "${PN}-core", "lib-dynload/gdbm.so" ) m.addPackage( "${PN}-image", "Python graphical image handling", "${PN}-core", "colorsys.* imghdr.* lib-dynload/imageop.so lib-dynload/rgbimg.so" ) m.addPackage( "${PN}-io", "Python low-level I/O", "${PN}-core ${PN}-math ${PN}-textutils", "lib-dynload/_socket.so lib-dynload/_io.so lib-dynload/_ssl.so lib-dynload/select.so lib-dynload/termios.so lib-dynload/cStringIO.so " + "pipes.* socket.* ssl.* tempfile.* StringIO.* io.* _pyio.*" ) m.addPackage( "${PN}-json", "Python JSON support", "${PN}-core ${PN}-math ${PN}-re", "json lib-dynload/_json.so" ) # package m.addPackage( "${PN}-lang", "Python low-level language support", "${PN}-core", "lib-dynload/_bisect.so lib-dynload/_collections.so lib-dynload/_heapq.so lib-dynload/_weakref.so lib-dynload/_functools.so " + "lib-dynload/array.so lib-dynload/itertools.so lib-dynload/operator.so lib-dynload/parser.so " + "atexit.* bisect.* code.* codeop.* collections.* dis.* functools.* heapq.* inspect.* keyword.* opcode.* symbol.* repr.* token.* " + "tokenize.* traceback.* weakref.*" ) m.addPackage( "${PN}-logging", "Python logging support", "${PN}-core ${PN}-io ${PN}-lang ${PN}-pickle ${PN}-stringold", "logging" ) # package m.addPackage( "${PN}-mailbox", "Python mailbox format support", "${PN}-core ${PN}-mime", "mailbox.*" ) m.addPackage( "${PN}-math", "Python math support", "${PN}-core ${PN}-crypt", "lib-dynload/cmath.so lib-dynload/math.so lib-dynload/_random.so random.* sets.*" ) m.addPackage( "${PN}-mime", "Python MIME handling APIs", "${PN}-core ${PN}-io", "mimetools.* uu.* quopri.* rfc822.* MimeWriter.*" ) m.addPackage( "${PN}-mmap", "Python memory-mapped file support", "${PN}-core ${PN}-io", "lib-dynload/mmap.so " ) m.addPackage( "${PN}-multiprocessing", "Python multiprocessing support", "${PN}-core ${PN}-io ${PN}-lang ${PN}-pickle ${PN}-threading ${PN}-ctypes ${PN}-mmap", "lib-dynload/_multiprocessing.so multiprocessing" ) # package m.addPackage( "${PN}-netclient", "Python Internet Protocol clients", "${PN}-core ${PN}-crypt ${PN}-datetime ${PN}-io ${PN}-lang ${PN}-logging ${PN}-mime", "*Cookie*.* " + "base64.* cookielib.* ftplib.* gopherlib.* hmac.* httplib.* mimetypes.* nntplib.* poplib.* smtplib.* telnetlib.* urllib.* urllib2.* urlparse.* uuid.* rfc822.* mimetools.*" ) m.addPackage( "${PN}-netserver", "Python Internet Protocol servers", "${PN}-core ${PN}-netclient", "cgi.* *HTTPServer.* SocketServer.*" ) m.addPackage( "${PN}-numbers", "Python number APIs", "${PN}-core ${PN}-lang ${PN}-re", "decimal.* numbers.*" ) m.addPackage( "${PN}-pickle", "Python serialisation/persistence support", "${PN}-core ${PN}-codecs ${PN}-io ${PN}-re", "pickle.* shelve.* lib-dynload/cPickle.so pickletools.*" ) m.addPackage( "${PN}-pkgutil", "Python package extension utility support", "${PN}-core", "pkgutil.*") m.addPackage( "${PN}-pprint", "Python pretty-print support", "${PN}-core ${PN}-io", "pprint.*" ) m.addPackage( "${PN}-profile", "Python basic performance profiling support", "${PN}-core ${PN}-textutils", "profile.* pstats.* cProfile.* lib-dynload/_lsprof.so" ) m.addPackage( "${PN}-re", "Python Regular Expression APIs", "${PN}-core", "re.* sre.* sre_compile.* sre_constants* sre_parse.*" ) # _sre is builtin m.addPackage( "${PN}-readline", "Python readline support", "${PN}-core", "lib-dynload/readline.so rlcompleter.*" ) m.addPackage( "${PN}-resource", "Python resource control interface", "${PN}-core", "lib-dynload/resource.so" ) m.addPackage( "${PN}-shell", "Python shell-like functionality", "${PN}-core ${PN}-re", "cmd.* commands.* dircache.* fnmatch.* glob.* popen2.* shlex.* shutil.*" ) m.addPackage( "${PN}-robotparser", "Python robots.txt parser", "${PN}-core ${PN}-netclient", "robotparser.*") m.addPackage( "${PN}-subprocess", "Python subprocess support", "${PN}-core ${PN}-io ${PN}-re ${PN}-fcntl ${PN}-pickle", "subprocess.*" ) m.addPackage( "${PN}-sqlite3", "Python Sqlite3 database support", "${PN}-core ${PN}-datetime ${PN}-lang ${PN}-crypt ${PN}-io ${PN}-threading ${PN}-zlib", "lib-dynload/_sqlite3.so sqlite3/dbapi2.* sqlite3/__init__.* sqlite3/dump.*" ) m.addPackage( "${PN}-sqlite3-tests", "Python Sqlite3 database support tests", "${PN}-core ${PN}-sqlite3", "sqlite3/test" ) m.addPackage( "${PN}-stringold", "Python string APIs [deprecated]", "${PN}-core ${PN}-re", "lib-dynload/strop.so string.* stringold.*" ) m.addPackage( "${PN}-syslog", "Python syslog interface", "${PN}-core", "lib-dynload/syslog.so" ) m.addPackage( "${PN}-terminal", "Python terminal controlling support", "${PN}-core ${PN}-io", "pty.* tty.*" ) m.addPackage( "${PN}-tests", "Python tests", "${PN}-core", "test" ) # package m.addPackage( "${PN}-threading", "Python threading & synchronization support", "${PN}-core ${PN}-lang", "_threading_local.* dummy_thread.* dummy_threading.* mutex.* threading.* Queue.*" ) m.addPackage( "${PN}-tkinter", "Python Tcl/Tk bindings", "${PN}-core", "lib-dynload/_tkinter.so lib-tk" ) # package m.addPackage( "${PN}-unittest", "Python unit testing framework", "${PN}-core ${PN}-stringold ${PN}-lang ${PN}-io ${PN}-difflib ${PN}-pprint ${PN}-shell", "unittest/" ) m.addPackage( "${PN}-unixadmin", "Python Unix administration support", "${PN}-core", "lib-dynload/nis.so lib-dynload/grp.so lib-dynload/pwd.so getpass.*" ) m.addPackage( "${PN}-xml", "Python basic XML support", "${PN}-core ${PN}-elementtree ${PN}-re", "lib-dynload/pyexpat.so xml xmllib.*" ) # package m.addPackage( "${PN}-xmlrpc", "Python XML-RPC support", "${PN}-core ${PN}-xml ${PN}-netserver ${PN}-lang", "xmlrpclib.* SimpleXMLRPCServer.* DocXMLRPCServer.*" ) m.addPackage( "${PN}-zlib", "Python zlib compression support", "${PN}-core", "lib-dynload/zlib.so" ) m.addPackage( "${PN}-mailbox", "Python mailbox format support", "${PN}-core ${PN}-mime", "mailbox.*" ) m.make() ########NEW FILE######## __FILENAME__ = generate-manifest-3.3 #!/usr/bin/env python # generate Python Manifest for the OpenEmbedded build system # (C) 2002-2010 Michael 'Mickey' Lauer <mlauer@vanille-media.de> # (C) 2007 Jeremy Laine # licensed under MIT, see COPYING.MIT # # June 22, 2011 -- Mark Hatle <mark.hatle@windriver.com> # * Updated to no longer generate special -dbg package, instead use the # single system -dbg # * Update version with ".1" to indicate this change # # 2014 Khem Raj <raj.khem@gmail.com> # Added python3 support # import os import sys import time VERSION = "3.3.3" __author__ = "Michael 'Mickey' Lauer <mlauer@vanille-media.de>" __version__ = "20140131" class MakefileMaker: def __init__( self, outfile ): """initialize""" self.packages = {} self.targetPrefix = "${libdir}/python%s/" % VERSION[:3] self.output = outfile self.out( """ # WARNING: This file is AUTO GENERATED: Manual edits will be lost next time I regenerate the file. # Generator: '%s' Version %s (C) 2002-2010 Michael 'Mickey' Lauer <mlauer@vanille-media.de> # Visit the Python for Embedded Systems Site => http://www.Vanille.de/projects/python.spy """ % ( sys.argv[0], __version__ ) ) # # helper functions # def out( self, data ): """print a line to the output file""" self.output.write( "%s\n" % data ) def setPrefix( self, targetPrefix ): """set a file prefix for addPackage files""" self.targetPrefix = targetPrefix def doProlog( self ): self.out( """ """ ) self.out( "" ) def addPackage( self, name, description, dependencies, filenames ): """add a package to the Makefile""" if type( filenames ) == type( "" ): filenames = filenames.split() fullFilenames = [] for filename in filenames: if filename[0] != "$": fullFilenames.append( "%s%s" % ( self.targetPrefix, filename ) ) else: fullFilenames.append( filename ) self.packages[name] = description, dependencies, fullFilenames def doBody( self ): """generate body of Makefile""" global VERSION # # generate provides line # provideLine = 'PROVIDES+="' for name in sorted(self.packages): provideLine += "%s " % name provideLine += '"' self.out( provideLine ) self.out( "" ) # # generate package line # packageLine = 'PACKAGES="${PN}-dbg ' for name in sorted(self.packages): if name.startswith("${PN}-distutils"): if name == "${PN}-distutils": packageLine += "%s-staticdev %s " % (name, name) elif name != '${PN}-dbg': packageLine += "%s " % name packageLine += '${PN}-modules"' self.out( packageLine ) self.out( "" ) # # generate package variables # for name, data in sorted(self.packages.iteritems()): desc, deps, files = data # # write out the description, revision and dependencies # self.out( 'SUMMARY_%s="%s"' % ( name, desc ) ) self.out( 'RDEPENDS_%s="%s"' % ( name, deps ) ) line = 'FILES_%s="' % name # # check which directories to make in the temporary directory # dirset = {} # if python had a set-datatype this would be sufficient. for now, we're using a dict instead. for target in files: dirset[os.path.dirname( target )] = True # # generate which files to copy for the target (-dfR because whole directories are also allowed) # for target in files: line += "%s " % target line += '"' self.out( line ) self.out( "" ) self.out( 'SUMMARY_${PN}-modules="All Python modules"' ) line = 'RDEPENDS_${PN}-modules="' for name, data in sorted(self.packages.iteritems()): if name not in ['${PN}-dev', '${PN}-distutils-staticdev']: line += "%s " % name self.out( "%s \"" % line ) self.out( 'ALLOW_EMPTY_${PN}-modules = "1"' ) def doEpilog( self ): self.out( """""" ) self.out( "" ) def make( self ): self.doProlog() self.doBody() self.doEpilog() if __name__ == "__main__": if len( sys.argv ) > 1: try: os.unlink(sys.argv[1]) except Exception: sys.exc_clear() outfile = file( sys.argv[1], "w" ) else: outfile = sys.stdout m = MakefileMaker( outfile ) # Add packages here. Only specify dlopen-style library dependencies here, no ldd-style dependencies! # Parameters: revision, name, description, dependencies, filenames # m.addPackage( "${PN}-core", "Python interpreter and core modules", "${PN}-lang ${PN}-re", "__future__.* _abcoll.* abc.* copy.* copy_reg.* ConfigParser.* " + "genericpath.* getopt.* linecache.* new.* " + "os.* posixpath.* struct.* " + "warnings.* site.* stat.* " + "UserDict.* UserList.* UserString.* " + "lib-dynload/binascii.*.so lib-dynload/_struct.*.so lib-dynload/time.*.so " + "lib-dynload/xreadlines.*.so types.* platform.* ${bindir}/python* " + "_weakrefset.* sysconfig.* config/Makefile " + "${includedir}/python${PYTHON_MAJMIN}/pyconfig*.h " + "${libdir}/python${PYTHON_MAJMIN}/collections " + "${libdir}/python${PYTHON_MAJMIN}/sitecustomize.py ") m.addPackage( "${PN}-dev", "Python development package", "${PN}-core", "${includedir} " + "${libdir}/lib*${SOLIBSDEV} " + "${libdir}/*.la " + "${libdir}/*.a " + "${libdir}/*.o " + "${libdir}/pkgconfig " + "${base_libdir}/*.a " + "${base_libdir}/*.o " + "${datadir}/aclocal " + "${datadir}/pkgconfig " ) m.addPackage( "${PN}-2to3", "Python automated Python 2 to 3 code translator", "${PN}-core", "${bindir}/2to3 lib2to3" ) # package m.addPackage( "${PN}-idle", "Python Integrated Development Environment", "${PN}-core ${PN}-tkinter", "${bindir}/idle idlelib" ) # package m.addPackage( "${PN}-pydoc", "Python interactive help support", "${PN}-core ${PN}-lang ${PN}-stringold ${PN}-re", "${bindir}/pydoc pydoc.* pydoc_data" ) m.addPackage( "${PN}-smtpd", "Python Simple Mail Transport Daemon", "${PN}-core ${PN}-netserver ${PN}-email ${PN}-mime", "${bindir}/smtpd.* smtpd.*" ) m.addPackage( "${PN}-audio", "Python Audio Handling", "${PN}-core", "wave.* chunk.* sndhdr.* lib-dynload/ossaudiodev.*.so lib-dynload/audioop.*.so audiodev.* sunaudio.* sunau.* toaiff.*" ) m.addPackage( "${PN}-codecs", "Python codecs, encodings & i18n support", "${PN}-core ${PN}-lang", "codecs.* encodings gettext.* locale.* lib-dynload/_locale.*.so lib-dynload/_codecs* lib-dynload/_multibytecodec.*.so lib-dynload/unicodedata.*.so stringprep.* xdrlib.*" ) m.addPackage( "${PN}-compile", "Python bytecode compilation support", "${PN}-core", "py_compile.* compileall.*" ) m.addPackage( "${PN}-compression", "Python high-level compression support", "${PN}-core ${PN}-codecs", "gzip.* zipfile.* tarfile.* lib-dynload/bz2.*.so" ) m.addPackage( "${PN}-crypt", "Python basic cryptographic and hashing support", "${PN}-core", "hashlib.* md5.* sha.* lib-dynload/crypt.*.so lib-dynload/_hashlib.*.so lib-dynload/_sha256.*.so lib-dynload/_sha512.*.so" ) m.addPackage( "${PN}-textutils", "Python option parsing, text wrapping and CSV support", "${PN}-core ${PN}-io ${PN}-re ${PN}-stringold", "lib-dynload/_csv.*.so csv.* optparse.* textwrap.*" ) m.addPackage( "${PN}-curses", "Python curses support", "${PN}-core", "curses lib-dynload/_curses.*.so lib-dynload/_curses_panel.*.so" ) # directory + low level module m.addPackage( "${PN}-ctypes", "Python C types support", "${PN}-core", "ctypes lib-dynload/_ctypes.*.so lib-dynload/_ctypes_test.*.so" ) # directory + low level module m.addPackage( "${PN}-datetime", "Python calendar and time support", "${PN}-core ${PN}-codecs", "_strptime.* calendar.* lib-dynload/datetime.*.so" ) m.addPackage( "${PN}-db", "Python file-based database support", "${PN}-core", "anydbm.* dumbdbm.* whichdb.* dbm lib-dynload/_dbm.*.so" ) m.addPackage( "${PN}-debugger", "Python debugger", "${PN}-core ${PN}-io ${PN}-lang ${PN}-re ${PN}-stringold ${PN}-shell ${PN}-pprint", "bdb.* pdb.*" ) m.addPackage( "${PN}-difflib", "Python helpers for computing deltas between objects", "${PN}-lang ${PN}-re", "difflib.*" ) m.addPackage( "${PN}-distutils-staticdev", "Python distribution utilities (static libraries)", "${PN}-distutils", "config/lib*.a" ) # package m.addPackage( "${PN}-distutils", "Python Distribution Utilities", "${PN}-core", "config distutils" ) # package m.addPackage( "${PN}-doctest", "Python framework for running examples in docstrings", "${PN}-core ${PN}-lang ${PN}-io ${PN}-re ${PN}-unittest ${PN}-debugger ${PN}-difflib", "doctest.*" ) # FIXME consider adding to some higher level package m.addPackage( "${PN}-elementtree", "Python elementree", "${PN}-core", "lib-dynload/_elementtree.*.so" ) m.addPackage( "${PN}-email", "Python email support", "${PN}-core ${PN}-io ${PN}-re ${PN}-mime ${PN}-audio ${PN}-image ${PN}-netclient", "imaplib.* email" ) # package m.addPackage( "${PN}-fcntl", "Python's fcntl interface", "${PN}-core", "lib-dynload/fcntl.*.so" ) m.addPackage( "${PN}-html", "Python HTML processing support", "${PN}-core", "formatter.* htmlentitydefs.* htmllib.* markupbase.* sgmllib.* HTMLParser.* " ) m.addPackage( "${PN}-gdbm", "Python GNU database support", "${PN}-core", "lib-dynload/_gdbm.*.so" ) m.addPackage( "${PN}-image", "Python graphical image handling", "${PN}-core", "colorsys.* imghdr.* lib-dynload/imageop.*.so lib-dynload/rgbimg.*.so" ) m.addPackage( "${PN}-io", "Python low-level I/O", "${PN}-core ${PN}-math", "lib-dynload/_socket.*.so lib-dynload/_io.*.so lib-dynload/_ssl.*.so lib-dynload/select.*.so lib-dynload/termios.*.so lib-dynload/cStringIO.*.so " + "pipes.* socket.* ssl.* tempfile.* StringIO.* io.* _pyio.*" ) m.addPackage( "${PN}-json", "Python JSON support", "${PN}-core ${PN}-math ${PN}-re", "json lib-dynload/_json.*.so" ) # package m.addPackage( "${PN}-lang", "Python low-level language support", "${PN}-core", "lib-dynload/_bisect.*.so lib-dynload/_collections.*.so lib-dynload/_heapq.*.so lib-dynload/_weakref.*.so lib-dynload/_functools.*.so " + "lib-dynload/array.*.so lib-dynload/itertools.*.so lib-dynload/operator.*.so lib-dynload/parser.*.so " + "atexit.* bisect.* code.* codeop.* collections.* dis.* functools.* heapq.* inspect.* keyword.* opcode.* symbol.* repr.* token.* " + "tokenize.* traceback.* weakref.*" ) m.addPackage( "${PN}-logging", "Python logging support", "${PN}-core ${PN}-io ${PN}-lang ${PN}-pickle ${PN}-stringold", "logging" ) # package m.addPackage( "${PN}-mailbox", "Python mailbox format support", "${PN}-core ${PN}-mime", "mailbox.*" ) m.addPackage( "${PN}-math", "Python math support", "${PN}-core", "lib-dynload/cmath.*.so lib-dynload/math.*.so lib-dynload/_random.*.so random.* sets.*" ) m.addPackage( "${PN}-mime", "Python MIME handling APIs", "${PN}-core ${PN}-io", "mimetools.* uu.* quopri.* rfc822.* MimeWriter.*" ) m.addPackage( "${PN}-mmap", "Python memory-mapped file support", "${PN}-core ${PN}-io", "lib-dynload/mmap.*.so " ) m.addPackage( "${PN}-multiprocessing", "Python multiprocessing support", "${PN}-core ${PN}-io ${PN}-lang ${PN}-pickle ${PN}-threading ${PN}-ctypes ${PN}-mmap", "lib-dynload/_multiprocessing.*.so multiprocessing" ) # package m.addPackage( "${PN}-netclient", "Python Internet Protocol clients", "${PN}-core ${PN}-crypt ${PN}-datetime ${PN}-io ${PN}-lang ${PN}-logging ${PN}-mime", "*Cookie*.* " + "base64.* cookielib.* ftplib.* gopherlib.* hmac.* httplib.* mimetypes.* nntplib.* poplib.* smtplib.* telnetlib.* urllib.* urllib2.* urlparse.* uuid.* rfc822.* mimetools.*" ) m.addPackage( "${PN}-netserver", "Python Internet Protocol servers", "${PN}-core ${PN}-netclient", "cgi.* *HTTPServer.* SocketServer.*" ) m.addPackage( "${PN}-numbers", "Python number APIs", "${PN}-core ${PN}-lang ${PN}-re", "decimal.* numbers.*" ) m.addPackage( "${PN}-pickle", "Python serialisation/persistence support", "${PN}-core ${PN}-codecs ${PN}-io ${PN}-re", "pickle.* shelve.* lib-dynload/cPickle.*.so pickletools.*" ) m.addPackage( "${PN}-pkgutil", "Python package extension utility support", "${PN}-core", "pkgutil.*") m.addPackage( "${PN}-pprint", "Python pretty-print support", "${PN}-core ${PN}-io", "pprint.*" ) m.addPackage( "${PN}-profile", "Python basic performance profiling support", "${PN}-core ${PN}-textutils", "profile.* pstats.* cProfile.* lib-dynload/_lsprof.*.so" ) m.addPackage( "${PN}-re", "Python Regular Expression APIs", "${PN}-core", "re.* sre.* sre_compile.* sre_constants* sre_parse.*" ) # _sre is builtin m.addPackage( "${PN}-readline", "Python readline support", "${PN}-core", "lib-dynload/readline.*.so rlcompleter.*" ) m.addPackage( "${PN}-resource", "Python resource control interface", "${PN}-core", "lib-dynload/resource.*.so" ) m.addPackage( "${PN}-shell", "Python shell-like functionality", "${PN}-core ${PN}-re", "cmd.* commands.* dircache.* fnmatch.* glob.* popen2.* shlex.* shutil.*" ) m.addPackage( "${PN}-robotparser", "Python robots.txt parser", "${PN}-core ${PN}-netclient", "urllib/robotparser.*") m.addPackage( "${PN}-subprocess", "Python subprocess support", "${PN}-core ${PN}-io ${PN}-re ${PN}-fcntl ${PN}-pickle", "subprocess.*" ) m.addPackage( "${PN}-sqlite3", "Python Sqlite3 database support", "${PN}-core ${PN}-datetime ${PN}-lang ${PN}-crypt ${PN}-io ${PN}-threading", "lib-dynload/_sqlite3.*.so sqlite3/dbapi2.* sqlite3/__init__.* sqlite3/dump.*" ) m.addPackage( "${PN}-sqlite3-tests", "Python Sqlite3 database support tests", "${PN}-core ${PN}-sqlite3", "sqlite3/test" ) m.addPackage( "${PN}-stringold", "Python string APIs [deprecated]", "${PN}-core ${PN}-re", "lib-dynload/strop.*.so string.* stringold.*" ) m.addPackage( "${PN}-syslog", "Python syslog interface", "${PN}-core", "lib-dynload/syslog.*.so" ) m.addPackage( "${PN}-terminal", "Python terminal controlling support", "${PN}-core ${PN}-io", "pty.* tty.*" ) m.addPackage( "${PN}-tests", "Python tests", "${PN}-core", "test" ) # package m.addPackage( "${PN}-threading", "Python threading & synchronization support", "${PN}-core ${PN}-lang", "_threading_local.* dummy_thread.* dummy_threading.* mutex.* threading.* Queue.*" ) m.addPackage( "${PN}-tkinter", "Python Tcl/Tk bindings", "${PN}-core", "lib-dynload/_tkinter.*.so lib-tk tkinter" ) # package m.addPackage( "${PN}-unittest", "Python unit testing framework", "${PN}-core ${PN}-stringold ${PN}-lang ${PN}-io ${PN}-difflib ${PN}-pprint ${PN}-shell", "unittest/" ) m.addPackage( "${PN}-unixadmin", "Python Unix administration support", "${PN}-core", "lib-dynload/nis.*.so lib-dynload/grp.*.so lib-dynload/pwd.*.so getpass.*" ) m.addPackage( "${PN}-xml", "Python basic XML support", "${PN}-core ${PN}-elementtree ${PN}-re", "lib-dynload/pyexpat.*.so xml xmllib.*" ) # package m.addPackage( "${PN}-xmlrpc", "Python XML-RPC support", "${PN}-core ${PN}-xml ${PN}-netserver ${PN}-lang", "xmlrpclib.* SimpleXMLRPCServer.* DocXMLRPCServer.* xmlrpc" ) m.addPackage( "${PN}-mailbox", "Python mailbox format support", "${PN}-core ${PN}-mime", "mailbox.*" ) m.make() ########NEW FILE######## __FILENAME__ = jhbuild2oe #!/usr/bin/env python # Available modulesets: # # bootstrap.modules # freedesktop.modules # gcj.modules # gnome-2.10.modules # gnome-2.12.modules # gnome-2.14.modules # gnome-2.16.modules # gnutls.modules # gtk28.modules # gtk.modules # xorg-7.0.modules # xorg.modules moduleset = 'xorg.modules' import cElementTree as ElementTree # import lxml.etree as ElementTree import re, os, bb, bb.data class Handlers(object): """ Class to act as a store for handlers of jhbuild xml elements, and as a dispatcher of parsed Elements to those handlers. These handlers exist to take an xml element from the jhbuild files and either produce bitbake metadata in self.packages, or produce data which will be used by other element handlers to do so. Handlers(filename) -> new object to parse and process jhbuild file of name 'filename'. """ cvsrootpat = re.compile(r''' \s* # Skip leading whitespace :(?P<scheme>[^:]+): # scheme (i.e. pserver, ext) ((?P<user>\S+?)@)? # username (?P<host>\S+?): # non-greedy match of the remote host (?P<path>\S+) # remote path ''', re.VERBOSE) def __init__(self, msfile): self.msfile = msfile self.msbasename = os.path.basename(msfile) self.msdirname = os.path.dirname(msfile) self.handled = {} self.cvsroots = {} self.repositories = {} self.packages = [] def handle(self, element, parent): import sys """ XML Element dispatch function. Can be called both from outside the Handlers object to initiate handling, and from within individual XML element handlers to ensure that dependent elements have been handled. Does not handle a given XML Element more than once, as it retains information about the handling state of the Elements it encounters. """ try: state = self.handled[element] except KeyError: pass except: return try: self.__class__.__dict__[element.tag](self, element, parent) self.handled[element] = True except KeyError: self.handled[element] = False sys.__stderr__.write('Unhandled element: %s\n' % element.tag) except Exception: sys.__stderr__.write('Error handling %s: %s:\n %s\n' % (element.tag, sys.exc_type, sys.exc_value)) self.handled[element] = False print('handle(%s, %s) -> %s' % (element, parent, self.handled[element])) return self.handled[element] def cvsroot(self, element, parent): # Rip apart the cvsroot style location to build a cvs:// url for # bitbake's usage in the cvsmodule handler. # root=":pserver:anoncvs@cvs.freedesktop.org:/cvs/fontconfig" print("cvsroot(%s, %s)" % (element, parent)) root = element.attrib.get('root') rootmatch = re.match(Handlers.cvsrootpat, root) name = element.attrib.get('name') user = rootmatch.group('user') or '' if user != '': pw = element.attrib.get('password') or '' if pw != '': pw = ':' + pw + '@' else: user = user + '@' print('user: %s' % user) print('pw: %s' % pw) host = rootmatch.group('host') print('host: %s' % host) path = rootmatch.group('path') or '/' print('path: %s' % path) root = "cvs://%s%s%s%s" % (user, pw, host, path) print('root: %s' % root) self.cvsroots[name] = root def cvsmodule(self, element, parent): rootlist = [root for root in list(parent) if root.attrib.get('name') == element.attrib.get('cvsroot')] if len(rootlist) < 1: raise Exception("Error: cvsmodule '%s' requires cvsroot '%s'." % (element.attrib.get('module'), element.attrib.get('cvsroot'))) cvsroot = rootlist[0] def include(self, element, parent): href = element.attrib.get('href') fullhref = os.path.join(self.msdirname, href) tree = ElementTree.ElementTree(file=fullhref) elem = tree.getroot() # Append the children of the newly included root element to the parent # element, and manually handle() them, as the currently running # iteration isn't going to hit them. for child in elem: self.handle(child, elem) parent.append(elem) def repository(self, element, parent): # TODO: # Convert the URL in the href attribute, if necessary, to the format # which bitbake expects to see in SRC_URI. name = element.attrib.get('name') self.repositories[name] = element.attrib.get('href') def moduleset(self, element, parent): for child in element: self.handle(child, element) def packagename(self, name): # mangle name into an appropriate bitbake package name return name.replace('/', '-') def metamodule(self, element, parent): # grab the deps deps = None for child in element: if child.tag == 'dependencies': deps = [self.packagename(dep.attrib.get('package')) for dep in child if dep.tag == "dep"] # create the package d = bb.data.init() pn = self.packagename(element.attrib.get('id')) d.setVar('PN', pn) bb.data.setVar('DEPENDS', ' '.join(deps), d) d.setVar('_handler', 'metamodule') self.packages.append(d) def autotools(self, element, parent): deps = None branch = None for child in element: if child.tag == 'dependencies': deps = [self.packagename(dep.attrib.get('package')) for dep in child if dep.tag == "dep"] elif child.tag == 'branch': branch = child # create the package d = bb.data.init() id = element.attrib.get('id') if id is None: raise Exception('Error: autotools element has no id attribute.') pn = self.packagename(id) d.setVar('PN', pn) if deps is not None: bb.data.setVar('DEPENDS', ' '.join(deps), d) if branch is not None: # <branch repo="git.freedesktop.org" module="xorg/xserver"/> repo = os.path.join(self.repositories[branch.attrib.get('repo')], branch.attrib.get('module')) d.setVar('SRC_URI', repo) checkoutdir = branch.attrib.get('checkoutdir') if checkoutdir is not None: bb.data.setVar('S', os.path.join('${WORKDIR}', checkoutdir), d) # build class d.setVar('INHERITS', 'autotools') d.setVarFlag('INHERITS', 'operator', '+=') d.setVar('_handler', 'autotools') self.packages.append(d) class Emitter(object): """ Class which contains a single method for the emission of a bitbake package from the bitbake data produced by a Handlers object. """ def __init__(self, filefunc = None, basedir = None): def _defaultfilefunc(package): # return a relative path to the bitbake .bb which will be written return package.getVar('PN', 1) + '.bb' self.filefunc = filefunc or _defaultfilefunc self.basedir = basedir or os.path.abspath(os.curdir) def write(self, package, template = None): # 1) Assemble new file contents in ram, either new from bitbake # metadata, or a combination of the template and that metadata. # 2) Open the path returned by the filefunc + the basedir for writing. # 3) Write the new bitbake data file. fdata = '' if template: f = file(template, 'r') fdata = f.read() f.close() for key in bb.data.keys(package): fdata = fdata.replace('@@'+key+'@@', package.getVar(key)) else: for key in bb.data.keys(package): if key == '_handler': continue elif key == 'INHERITS': fdata += 'inherit %s\n' % package.getVar('INHERITS') else: oper = package.getVarFlag(key, 'operator') or '=' fdata += '%s %s "%s"\n' % (key, oper, package.getVar(key)) if not os.path.exists(os.path.join(self.basedir, os.path.dirname(self.filefunc(package)))): os.makedirs(os.path.join(self.basedir, os.path.dirname(self.filefunc(package)))) out = file(os.path.join(self.basedir, self.filefunc(package)), 'w') out.write(fdata) out.close() def _test(): msfile = os.path.join(os.path.abspath(os.curdir), 'modulesets', moduleset) tree = ElementTree.ElementTree(file=msfile) elem = tree.getroot() handlers = Handlers(msfile) handlers.handle(elem, None) def filefunc(package): # return a relative path to the bitbake .bb which will be written src_uri = package.getVar('SRC_URI', 1) filename = package.getVar('PN', 1) + '.bb' if not src_uri: return filename else: substr = src_uri[src_uri.find('xorg/'):] subdirlist = substr.split('/')[:2] subdir = '-'.join(subdirlist) return os.path.join(subdir, filename) emitter = Emitter(filefunc) for package in handlers.packages: template = emitter.filefunc(package) + '.in' if os.path.exists(template): print("%s exists, emitting based on template" % template) emitter.write(package, template) else: print("%s does not exist, emitting non-templated" % template) emitter.write(package) if __name__ == "__main__": _test() ########NEW FILE######## __FILENAME__ = engine # ex:ts=4:sw=4:sts=4:et # -*- tab-width: 4; c-basic-offset: 4; indent-tabs-mode: nil -*- # # Copyright (c) 2013, Intel Corporation. # All rights reserved. # # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License version 2 as # published by the Free Software Foundation. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., # 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA. # # DESCRIPTION # This module implements the image creation engine used by 'wic' to # create images. The engine parses through the OpenEmbedded kickstart # (wks) file specified and generates images that can then be directly # written onto media. # # AUTHORS # Tom Zanussi <tom.zanussi (at] linux.intel.com> # import os import sys from abc import ABCMeta, abstractmethod import shlex import json import subprocess import shutil import os, sys, errno from mic import msger, creator from mic.utils import cmdln, misc, errors from mic.conf import configmgr from mic.plugin import pluginmgr from mic.__version__ import VERSION from mic.utils.oe.misc import * def verify_build_env(): """ Verify that the build environment is sane. Returns True if it is, false otherwise """ try: builddir = os.environ["BUILDDIR"] except KeyError: print "BUILDDIR not found, exiting. (Did you forget to source oe-init-build-env?)" sys.exit(1) return True def find_artifacts(image_name): """ Gather the build artifacts for the current image (the image_name e.g. core-image-minimal) for the current MACHINE set in local.conf """ bitbake_env_lines = get_bitbake_env_lines() rootfs_dir = kernel_dir = hdddir = staging_data_dir = native_sysroot = "" for line in bitbake_env_lines.split('\n'): if (get_line_val(line, "IMAGE_ROOTFS")): rootfs_dir = get_line_val(line, "IMAGE_ROOTFS") continue if (get_line_val(line, "STAGING_KERNEL_DIR")): kernel_dir = get_line_val(line, "STAGING_KERNEL_DIR") continue if (get_line_val(line, "HDDDIR")): hdddir = get_line_val(line, "HDDDIR") continue if (get_line_val(line, "STAGING_DATADIR")): staging_data_dir = get_line_val(line, "STAGING_DATADIR") continue if (get_line_val(line, "STAGING_DIR_NATIVE")): native_sysroot = get_line_val(line, "STAGING_DIR_NATIVE") continue return (rootfs_dir, kernel_dir, hdddir, staging_data_dir, native_sysroot) CANNED_IMAGE_DIR = "lib/image/canned-wks" # relative to scripts SCRIPTS_CANNED_IMAGE_DIR = "scripts/" + CANNED_IMAGE_DIR def build_canned_image_list(dl): layers_path = get_bitbake_var("BBLAYERS") canned_wks_layer_dirs = [] for layer_path in layers_path.split(): path = os.path.join(layer_path, SCRIPTS_CANNED_IMAGE_DIR) canned_wks_layer_dirs.append(path) path = os.path.join(dl, CANNED_IMAGE_DIR) canned_wks_layer_dirs.append(path) return canned_wks_layer_dirs def find_canned_image(scripts_path, wks_file): """ Find a .wks file with the given name in the canned files dir. Return False if not found """ layers_canned_wks_dir = build_canned_image_list(scripts_path) for canned_wks_dir in layers_canned_wks_dir: for root, dirs, files in os.walk(canned_wks_dir): for file in files: if file.endswith("~") or file.endswith("#"): continue if file.endswith(".wks") and wks_file + ".wks" == file: fullpath = os.path.join(canned_wks_dir, file) return fullpath return None def list_canned_images(scripts_path): """ List the .wks files in the canned image dir, minus the extension. """ layers_canned_wks_dir = build_canned_image_list(scripts_path) for canned_wks_dir in layers_canned_wks_dir: for root, dirs, files in os.walk(canned_wks_dir): for file in files: if file.endswith("~") or file.endswith("#"): continue if file.endswith(".wks"): fullpath = os.path.join(canned_wks_dir, file) f = open(fullpath, "r") lines = f.readlines() for line in lines: desc = "" idx = line.find("short-description:") if idx != -1: desc = line[idx + len("short-description:"):].strip() break basename = os.path.splitext(file)[0] print " %s\t\t%s" % (basename.ljust(30), desc) def list_canned_image_help(scripts_path, fullpath): """ List the help and params in the specified canned image. """ f = open(fullpath, "r") lines = f.readlines() found = False for line in lines: if not found: idx = line.find("long-description:") if idx != -1: print print line[idx + len("long-description:"):].strip() found = True continue if not line.strip(): break idx = line.find("#") if idx != -1: print line[idx + len("#:"):].rstrip() else: break def wic_create(args, wks_file, rootfs_dir, bootimg_dir, kernel_dir, native_sysroot, hdddir, staging_data_dir, scripts_path, image_output_dir, debug, properties_file, properties=None): """ Create image wks_file - user-defined OE kickstart file rootfs_dir - absolute path to the build's /rootfs dir bootimg_dir - absolute path to the build's boot artifacts directory kernel_dir - absolute path to the build's kernel directory native_sysroot - absolute path to the build's native sysroots dir hdddir - absolute path to the build's HDDDIR dir staging_data_dir - absolute path to the build's STAGING_DATA_DIR dir scripts_path - absolute path to /scripts dir image_output_dir - dirname to create for image properties_file - use values from this file if nonempty i.e no prompting properties - use values from this string if nonempty i.e no prompting Normally, the values for the build artifacts values are determined by 'wic -e' from the output of the 'bitbake -e' command given an image name e.g. 'core-image-minimal' and a given machine set in local.conf. If that's the case, the variables get the following values from the output of 'bitbake -e': rootfs_dir: IMAGE_ROOTFS kernel_dir: STAGING_KERNEL_DIR native_sysroot: STAGING_DIR_NATIVE hdddir: HDDDIR staging_data_dir: STAGING_DATA_DIR In the above case, bootimg_dir remains unset and the image creation code determines which of the passed-in directories to use. In the case where the values are passed in explicitly i.e 'wic -e' is not used but rather the individual 'wic' options are used to explicitly specify these values, hdddir and staging_data_dir will be unset, but bootimg_dir must be explicit i.e. explicitly set to either hdddir or staging_data_dir, depending on the image being generated. The other values (rootfs_dir, kernel_dir, and native_sysroot) correspond to the same values found above via 'bitbake -e'). """ try: oe_builddir = os.environ["BUILDDIR"] except KeyError: print "BUILDDIR not found, exiting. (Did you forget to source oe-init-build-env?)" sys.exit(1) direct_args = list() direct_args.insert(0, oe_builddir) direct_args.insert(0, image_output_dir) direct_args.insert(0, wks_file) direct_args.insert(0, rootfs_dir) direct_args.insert(0, bootimg_dir) direct_args.insert(0, kernel_dir) direct_args.insert(0, native_sysroot) direct_args.insert(0, hdddir) direct_args.insert(0, staging_data_dir) direct_args.insert(0, "direct") if debug: msger.set_loglevel('debug') cr = creator.Creator() cr.main(direct_args) print "\nThe image(s) were created using OE kickstart file:\n %s" % wks_file def wic_list(args, scripts_path, properties_file): """ Print the complete list of properties defined by the image, or the possible values for a particular image property. """ if len(args) < 1: return False if len(args) == 1: if args[0] == "images": list_canned_images(scripts_path) return True elif args[0] == "properties": return True else: return False if len(args) == 2: if args[0] == "properties": wks_file = args[1] print "print properties contained in wks file: %s" % wks_file return True elif args[0] == "property": print "print property values for property: %s" % args[1] return True elif args[1] == "help": wks_file = args[0] fullpath = find_canned_image(scripts_path, wks_file) if not fullpath: print "No image named %s found, exiting. (Use 'wic list images' to list available images, or specify a fully-qualified OE kickstart (.wks) filename)\n" % wks_file sys.exit(1) list_canned_image_help(scripts_path, fullpath) return True else: return False return False ########NEW FILE######## __FILENAME__ = help # ex:ts=4:sw=4:sts=4:et # -*- tab-width: 4; c-basic-offset: 4; indent-tabs-mode: nil -*- # # Copyright (c) 2013, Intel Corporation. # All rights reserved. # # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License version 2 as # published by the Free Software Foundation. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., # 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA. # # DESCRIPTION # This module implements some basic help invocation functions along # with the bulk of the help topic text for the OE Core Image Tools. # # AUTHORS # Tom Zanussi <tom.zanussi (at] linux.intel.com> # import subprocess import logging def subcommand_error(args): logging.info("invalid subcommand %s" % args[0]) def display_help(subcommand, subcommands): """ Display help for subcommand. """ if subcommand not in subcommands: return False help = subcommands.get(subcommand, subcommand_error)[2] pager = subprocess.Popen('less', stdin=subprocess.PIPE) pager.communicate(help) return True def wic_help(args, usage_str, subcommands): """ Subcommand help dispatcher. """ if len(args) == 1 or not display_help(args[1], subcommands): print(usage_str) def invoke_subcommand(args, parser, main_command_usage, subcommands): """ Dispatch to subcommand handler borrowed from combo-layer. Should use argparse, but has to work in 2.6. """ if not args: logging.error("No subcommand specified, exiting") parser.print_help() elif args[0] == "help": wic_help(args, main_command_usage, subcommands) elif args[0] not in subcommands: logging.error("Unsupported subcommand %s, exiting\n" % (args[0])) parser.print_help() else: usage = subcommands.get(args[0], subcommand_error)[1] subcommands.get(args[0], subcommand_error)[0](args[1:], usage) ## # wic help and usage strings ## wic_usage = """ Create a customized OpenEmbedded image usage: wic [--version] [--help] COMMAND [ARGS] Current 'wic' commands are: create Create a new OpenEmbedded image list List available values for options and image properties See 'wic help COMMAND' for more information on a specific command. """ wic_help_usage = """ usage: wic help <subcommand> This command displays detailed help for the specified subcommand. """ wic_create_usage = """ Create a new OpenEmbedded image usage: wic create <wks file or image name> [-o <DIRNAME> | --outdir <DIRNAME>] [-i <JSON PROPERTY FILE> | --infile <JSON PROPERTY_FILE>] [-e | --image-name] [-r, --rootfs-dir] [-b, --bootimg-dir] [-k, --kernel-dir] [-n, --native-sysroot] [-s, --skip-build-check] This command creates an OpenEmbedded image based on the 'OE kickstart commands' found in the <wks file>. The -o option can be used to place the image in a directory with a different name and location. See 'wic help create' for more detailed instructions. """ wic_create_help = """ NAME wic create - Create a new OpenEmbedded image SYNOPSIS wic create <wks file or image name> [-o <DIRNAME> | --outdir <DIRNAME>] [-i <JSON PROPERTY FILE> | --infile <JSON PROPERTY_FILE>] [-e | --image-name] [-r, --rootfs-dir] [-b, --bootimg-dir] [-k, --kernel-dir] [-n, --native-sysroot] [-s, --skip-build-check] DESCRIPTION This command creates an OpenEmbedded image based on the 'OE kickstart commands' found in the <wks file>. In order to do this, wic needs to know the locations of the various build artifacts required to build the image. Users can explicitly specify the build artifact locations using the -r, -b, -k, and -n options. See below for details on where the corresponding artifacts are typically found in a normal OpenEmbedded build. Alternatively, users can use the -e option to have 'mic' determine those locations for a given image. If the -e option is used, the user needs to have set the appropriate MACHINE variable in local.conf, and have sourced the build environment. The -e option is used to specify the name of the image to use the artifacts from e.g. core-image-sato. The -r option is used to specify the path to the /rootfs dir to use as the .wks rootfs source. The -b option is used to specify the path to the dir containing the boot artifacts (e.g. /EFI or /syslinux dirs) to use as the .wks bootimg source. The -k option is used to specify the path to the dir containing the kernel to use in the .wks bootimg. The -n option is used to specify the path to the native sysroot containing the tools to use to build the image. The -s option is used to skip the build check. The build check is a simple sanity check used to determine whether the user has sourced the build environment so that the -e option can operate correctly. If the user has specified the build artifact locations explicitly, 'wic' assumes the user knows what he or she is doing and skips the build check. When 'wic -e' is used, the locations for the build artifacts values are determined by 'wic -e' from the output of the 'bitbake -e' command given an image name e.g. 'core-image-minimal' and a given machine set in local.conf. In that case, the image is created as if the following 'bitbake -e' variables were used: -r: IMAGE_ROOTFS -k: STAGING_KERNEL_DIR -n: STAGING_DIR_NATIVE -b: HDDDIR and STAGING_DATA_DIR (handlers decide which to use) If 'wic -e' is not used, the user needs to select the appropriate value for -b (as well as -r, -k, and -n). The -o option can be used to place the image in a directory with a different name and location. As an alternative to the wks file, the image-specific properties that define the values that will be used to generate a particular image can be specified on the command-line using the -i option and supplying a JSON object consisting of the set of name:value pairs needed by image creation. The set of properties available for a given image type can be listed using the 'wic list' command. """ wic_list_usage = """ List available OpenEmbedded image properties and values usage: wic list images wic list <image> help wic list properties wic list properties <wks file> wic list property <property> [-o <JSON PROPERTY FILE> | --outfile <JSON PROPERTY_FILE>] This command enumerates the set of available canned images as well as help for those images. It also can be used to enumerate the complete set of possible values for a specified option or property needed by the image creation process. The first form enumerates all the available 'canned' images. The second form lists the detailed help information for a specific 'canned' image. The third form enumerates all the possible values that exist and can be specified in an OE kickstart (wks) file. The fourth form enumerates all the possible options that exist for the set of properties specified in a given OE kickstart (ks) file. The final form enumerates all the possible values that exist and can be specified for any given OE kickstart (wks) property. See 'wic help list' for more details. """ wic_list_help = """ NAME wic list - List available OpenEmbedded image properties and values SYNOPSIS wic list images wic list <image> help wic list properties wic list properties <wks file> wic list property <property> [-o <JSON PROPERTY FILE> | --outfile <JSON PROPERTY_FILE>] DESCRIPTION This command enumerates the complete set of possible values for a specified option or property needed by the image creation process. This command enumerates the set of available canned images as well as help for those images. It also can be used to enumerate the complete set of possible values for a specified option or property needed by the image creation process. The first form enumerates all the available 'canned' images. These are actually just the set of .wks files that have been moved into the /scripts/lib/image/canned-wks directory). The second form lists the detailed help information for a specific 'canned' image. The third form enumerates all the possible values that exist and can be specified in a OE kickstart (wks) file. The output of this can be used by the third form to print the description and possible values of a specific property. The fourth form enumerates all the possible options that exist for the set of properties specified in a given OE kickstart (wks) file. If the -o option is specified, the list of properties, in addition to being displayed, will be written to the specified file as a JSON object. In this case, the object will consist of the set of name:value pairs corresponding to the (possibly nested) dictionary of properties defined by the input statements used by the image. Some example output for the 'list <wks file>' command: $ wic list test.ks "part" : { "mountpoint" : "/" "fstype" : "ext3" } "part" : { "mountpoint" : "/home" "fstype" : "ext3" "offset" : "10000" } "bootloader" : { "type" : "efi" } . . . Each entry in the output consists of the name of the input element e.g. "part", followed by the properties defined for that element enclosed in braces. This information should provide sufficient information to create a complete user interface with. The final form enumerates all the possible values that exist and can be specified for any given OE kickstart (wks) property. If the -o option is specified, the list of values for the given property, in addition to being displayed, will be written to the specified file as a JSON object. In this case, the object will consist of the set of name:value pairs corresponding to the array of property values associated with the property. $ wic list property part ["mountpoint", "where the partition should be mounted"] ["fstype", "filesytem type of the partition"] ["ext3"] ["ext4"] ["btrfs"] ["swap"] ["offset", "offset of the partition within the image"] """ ########NEW FILE######## __FILENAME__ = base # # Chris Lumens <clumens@redhat.com> # # Copyright 2006, 2007, 2008 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # """ Base classes for creating commands and syntax version object. This module exports several important base classes: BaseData - The base abstract class for all data objects. Data objects are contained within a BaseHandler object. BaseHandler - The base abstract class from which versioned kickstart handler are derived. Subclasses of BaseHandler hold BaseData and KickstartCommand objects. DeprecatedCommand - An abstract subclass of KickstartCommand that should be further subclassed by users of this module. When a subclass is used, a warning message will be printed. KickstartCommand - The base abstract class for all kickstart commands. Command objects are contained within a BaseHandler object. """ import gettext gettext.textdomain("pykickstart") _ = lambda x: gettext.ldgettext("pykickstart", x) import types import warnings from pykickstart.errors import * from pykickstart.ko import * from pykickstart.parser import Packages from pykickstart.version import versionToString ### ### COMMANDS ### class KickstartCommand(KickstartObject): """The base class for all kickstart commands. This is an abstract class.""" removedKeywords = [] removedAttrs = [] def __init__(self, writePriority=0, *args, **kwargs): """Create a new KickstartCommand instance. This method must be provided by all subclasses, but subclasses must call KickstartCommand.__init__ first. Instance attributes: currentCmd -- The name of the command in the input file that caused this handler to be run. currentLine -- The current unprocessed line from the input file that caused this handler to be run. handler -- A reference to the BaseHandler subclass this command is contained withing. This is needed to allow referencing of Data objects. lineno -- The current line number in the input file. writePriority -- An integer specifying when this command should be printed when iterating over all commands' __str__ methods. The higher the number, the later this command will be written. All commands with the same priority will be written alphabetically. """ # We don't want people using this class by itself. if self.__class__ is KickstartCommand: raise TypeError, "KickstartCommand is an abstract class." KickstartObject.__init__(self, *args, **kwargs) self.writePriority = writePriority # These will be set by the dispatcher. self.currentCmd = "" self.currentLine = "" self.handler = None self.lineno = 0 # If a subclass provides a removedKeywords list, remove all the # members from the kwargs list before we start processing it. This # ensures that subclasses don't continue to recognize arguments that # were removed. for arg in filter(kwargs.has_key, self.removedKeywords): kwargs.pop(arg) def __call__(self, *args, **kwargs): """Set multiple attributes on a subclass of KickstartCommand at once via keyword arguments. Valid attributes are anything specified in a subclass, but unknown attributes will be ignored. """ for (key, val) in kwargs.items(): # Ignore setting attributes that were removed in a subclass, as # if they were unknown attributes. if key in self.removedAttrs: continue if hasattr(self, key): setattr(self, key, val) def __str__(self): """Return a string formatted for output to a kickstart file. This method must be provided by all subclasses. """ return KickstartObject.__str__(self) def parse(self, args): """Parse the list of args and set data on the KickstartCommand object. This method must be provided by all subclasses. """ raise TypeError, "parse() not implemented for KickstartCommand" def apply(self, instroot="/"): """Write out the configuration related to the KickstartCommand object. Subclasses which do not provide this method will not have their configuration written out. """ return def dataList(self): """For commands that can occur multiple times in a single kickstart file (like network, part, etc.), return the list that we should append more data objects to. """ return None def deleteRemovedAttrs(self): """Remove all attributes from self that are given in the removedAttrs list. This method should be called from __init__ in a subclass, but only after the superclass's __init__ method has been called. """ for attr in filter(lambda k: hasattr(self, k), self.removedAttrs): delattr(self, attr) # Set the contents of the opts object (an instance of optparse.Values # returned by parse_args) as attributes on the KickstartCommand object. # It's useful to call this from KickstartCommand subclasses after parsing # the arguments. def _setToSelf(self, optParser, opts): self._setToObj(optParser, opts, self) # Sets the contents of the opts object (an instance of optparse.Values # returned by parse_args) as attributes on the provided object obj. It's # useful to call this from KickstartCommand subclasses that handle lists # of objects (like partitions, network devices, etc.) and need to populate # a Data object. def _setToObj(self, optParser, opts, obj): for key in filter (lambda k: getattr(opts, k) != None, optParser.keys()): setattr(obj, key, getattr(opts, key)) class DeprecatedCommand(KickstartCommand): """Specify that a command is deprecated and no longer has any function. Any command that is deprecated should be subclassed from this class, only specifying an __init__ method that calls the superclass's __init__. This is an abstract class. """ def __init__(self, writePriority=None, *args, **kwargs): # We don't want people using this class by itself. if self.__class__ is KickstartCommand: raise TypeError, "DeprecatedCommand is an abstract class." # Create a new DeprecatedCommand instance. KickstartCommand.__init__(self, writePriority, *args, **kwargs) def __str__(self): """Placeholder since DeprecatedCommands don't work anymore.""" return "" def parse(self, args): """Print a warning message if the command is seen in the input file.""" mapping = {"lineno": self.lineno, "cmd": self.currentCmd} warnings.warn(_("Ignoring deprecated command on line %(lineno)s: The %(cmd)s command has been deprecated and no longer has any effect. It may be removed from future releases, which will result in a fatal error from kickstart. Please modify your kickstart file to remove this command.") % mapping, DeprecationWarning) ### ### HANDLERS ### class BaseHandler(KickstartObject): """Each version of kickstart syntax is provided by a subclass of this class. These subclasses are what users will interact with for parsing, extracting data, and writing out kickstart files. This is an abstract class. version -- The version this syntax handler supports. This is set by a class attribute of a BaseHandler subclass and is used to set up the command dict. It is for read-only use. """ version = None def __init__(self, mapping=None, dataMapping=None, commandUpdates=None, dataUpdates=None, *args, **kwargs): """Create a new BaseHandler instance. This method must be provided by all subclasses, but subclasses must call BaseHandler.__init__ first. mapping -- A custom map from command strings to classes, useful when creating your own handler with special command objects. It is otherwise unused and rarely needed. If you give this argument, the mapping takes the place of the default one and so must include all commands you want recognized. dataMapping -- This is the same as mapping, but for data objects. All the same comments apply. commandUpdates -- This is similar to mapping, but does not take the place of the defaults entirely. Instead, this mapping is applied after the defaults and updates it with just the commands you want to modify. dataUpdates -- This is the same as commandUpdates, but for data objects. Instance attributes: commands -- A mapping from a string command to a KickstartCommand subclass object that handles it. Multiple strings can map to the same object, but only one instance of the command object should ever exist. Most users should never have to deal with this directly, as it is manipulated internally and called through dispatcher. currentLine -- The current unprocessed line from the input file that caused this handler to be run. packages -- An instance of pykickstart.parser.Packages which describes the packages section of the input file. platform -- A string describing the hardware platform, which is needed only by system-config-kickstart. scripts -- A list of pykickstart.parser.Script instances, which is populated by KickstartParser.addScript and describes the %pre/%post/%traceback script section of the input file. """ # We don't want people using this class by itself. if self.__class__ is BaseHandler: raise TypeError, "BaseHandler is an abstract class." KickstartObject.__init__(self, *args, **kwargs) # This isn't really a good place for these, but it's better than # everything else I can think of. self.scripts = [] self.packages = Packages() self.platform = "" # These will be set by the dispatcher. self.commands = {} self.currentLine = 0 # A dict keyed by an integer priority number, with each value being a # list of KickstartCommand subclasses. This dict is maintained by # registerCommand and used in __str__. No one else should be touching # it. self._writeOrder = {} self._registerCommands(mapping, dataMapping, commandUpdates, dataUpdates) def __str__(self): """Return a string formatted for output to a kickstart file.""" retval = "" if self.platform != "": retval += "#platform=%s\n" % self.platform retval += "#version=%s\n" % versionToString(self.version) lst = self._writeOrder.keys() lst.sort() for prio in lst: for obj in self._writeOrder[prio]: retval += obj.__str__() for script in self.scripts: retval += script.__str__() retval += self.packages.__str__() return retval def _insertSorted(self, lst, obj): length = len(lst) i = 0 while i < length: # If the two classes have the same name, it's because we are # overriding an existing class with one from a later kickstart # version, so remove the old one in favor of the new one. if obj.__class__.__name__ > lst[i].__class__.__name__: i += 1 elif obj.__class__.__name__ == lst[i].__class__.__name__: lst[i] = obj return elif obj.__class__.__name__ < lst[i].__class__.__name__: break if i >= length: lst.append(obj) else: lst.insert(i, obj) def _setCommand(self, cmdObj): # Add an attribute on this version object. We need this to provide a # way for clients to access the command objects. We also need to strip # off the version part from the front of the name. if cmdObj.__class__.__name__.find("_") != -1: name = unicode(cmdObj.__class__.__name__.split("_", 1)[1]) else: name = unicode(cmdObj.__class__.__name__).lower() setattr(self, name.lower(), cmdObj) # Also, add the object into the _writeOrder dict in the right place. if cmdObj.writePriority is not None: if self._writeOrder.has_key(cmdObj.writePriority): self._insertSorted(self._writeOrder[cmdObj.writePriority], cmdObj) else: self._writeOrder[cmdObj.writePriority] = [cmdObj] def _registerCommands(self, mapping=None, dataMapping=None, commandUpdates=None, dataUpdates=None): if mapping == {} or mapping == None: from pykickstart.handlers.control import commandMap cMap = commandMap[self.version] else: cMap = mapping if dataMapping == {} or dataMapping == None: from pykickstart.handlers.control import dataMap dMap = dataMap[self.version] else: dMap = dataMapping if type(commandUpdates) == types.DictType: cMap.update(commandUpdates) if type(dataUpdates) == types.DictType: dMap.update(dataUpdates) for (cmdName, cmdClass) in cMap.iteritems(): # First make sure we haven't instantiated this command handler # already. If we have, we just need to make another mapping to # it in self.commands. cmdObj = None for (key, val) in self.commands.iteritems(): if val.__class__.__name__ == cmdClass.__name__: cmdObj = val break # If we didn't find an instance in self.commands, create one now. if cmdObj == None: cmdObj = cmdClass() self._setCommand(cmdObj) # Finally, add the mapping to the commands dict. self.commands[cmdName] = cmdObj self.commands[cmdName].handler = self # We also need to create attributes for the various data objects. # No checks here because dMap is a bijection. At least, that's what # the comment says. Hope no one screws that up. for (dataName, dataClass) in dMap.iteritems(): setattr(self, dataName, dataClass) def dispatcher(self, args, lineno): """Call the appropriate KickstartCommand handler for the current line in the kickstart file. A handler for the current command should be registered, though a handler of None is not an error. Returns the data object returned by KickstartCommand.parse. args -- A list of arguments to the current command lineno -- The line number in the file, for error reporting """ cmd = args[0] if not self.commands.has_key(cmd): raise KickstartParseError, formatErrorMsg(lineno, msg=_("Unknown command: %s" % cmd)) elif self.commands[cmd] != None: self.commands[cmd].currentCmd = cmd self.commands[cmd].currentLine = self.currentLine self.commands[cmd].lineno = lineno # The parser returns the data object that was modified. This could # be a BaseData subclass that should be put into a list, or it # could be the command handler object itself. obj = self.commands[cmd].parse(args[1:]) lst = self.commands[cmd].dataList() if lst is not None: lst.append(obj) return obj def maskAllExcept(self, lst): """Set all entries in the commands dict to None, except the ones in the lst. All other commands will not be processed. """ self._writeOrder = {} for (key, val) in self.commands.iteritems(): if not key in lst: self.commands[key] = None def hasCommand(self, cmd): """Return true if there is a handler for the string cmd.""" return hasattr(self, cmd) ### ### DATA ### class BaseData(KickstartObject): """The base class for all data objects. This is an abstract class.""" removedKeywords = [] removedAttrs = [] def __init__(self, *args, **kwargs): """Create a new BaseData instance. lineno -- Line number in the ks-file where this object was defined """ # We don't want people using this class by itself. if self.__class__ is BaseData: raise TypeError, "BaseData is an abstract class." KickstartObject.__init__(self, *args, **kwargs) self.lineno = 0 def __str__(self): """Return a string formatted for output to a kickstart file.""" return "" def __call__(self, *args, **kwargs): """Set multiple attributes on a subclass of BaseData at once via keyword arguments. Valid attributes are anything specified in a subclass, but unknown attributes will be ignored. """ for (key, val) in kwargs.items(): # Ignore setting attributes that were removed in a subclass, as # if they were unknown attributes. if key in self.removedAttrs: continue if hasattr(self, key): setattr(self, key, val) def deleteRemovedAttrs(self): """Remove all attributes from self that are given in the removedAttrs list. This method should be called from __init__ in a subclass, but only after the superclass's __init__ method has been called. """ for attr in filter(lambda k: hasattr(self, k), self.removedAttrs): delattr(self, attr) ########NEW FILE######## __FILENAME__ = authconfig # # Chris Lumens <clumens@redhat.com> # # Copyright 2005, 2006, 2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * class FC3_Authconfig(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): KickstartCommand.__init__(self, *args, **kwargs) self.authconfig = kwargs.get("authconfig", "") def __str__(self): retval = KickstartCommand.__str__(self) if self.authconfig: retval += "# System authorization information\nauth %s\n" % self.authconfig return retval def parse(self, args): self.authconfig = self.currentLine[len(self.currentCmd):].strip() return self ########NEW FILE######## __FILENAME__ = autopart # # Chris Lumens <clumens@redhat.com> # # Copyright 2005, 2006, 2007, 2008 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.errors import * from pykickstart.options import * import gettext _ = lambda x: gettext.ldgettext("pykickstart", x) class FC3_AutoPart(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=100, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.autopart = kwargs.get("autopart", False) def __str__(self): retval = KickstartCommand.__str__(self) if self.autopart: retval += "autopart\n" return retval def parse(self, args): if len(args) > 0: raise KickstartValueError, formatErrorMsg(self.lineno, msg=_("Kickstart command %s does not take any arguments") % "autopart") self.autopart = True return self class F9_AutoPart(FC3_AutoPart): removedKeywords = FC3_AutoPart.removedKeywords removedAttrs = FC3_AutoPart.removedAttrs def __init__(self, writePriority=100, *args, **kwargs): FC3_AutoPart.__init__(self, writePriority=writePriority, *args, **kwargs) self.encrypted = kwargs.get("encrypted", False) self.passphrase = kwargs.get("passphrase", "") self.op = self._getParser() def __str__(self): retval = KickstartCommand.__str__(self) if self.autopart: retval += "autopart" if self.encrypted: retval += " --encrypted" if self.passphrase != "": retval += " --passphrase=\"%s\""% self.passphrase if retval != "": retval += "\n" return retval def _getParser(self): op = KSOptionParser() op.add_option("--encrypted", action="store_true", default=False) op.add_option("--passphrase") return op def parse(self, args): (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) self._setToSelf(self.op, opts) self.autopart = True return self class F12_AutoPart(F9_AutoPart): removedKeywords = F9_AutoPart.removedKeywords removedAttrs = F9_AutoPart.removedAttrs def __init__(self, writePriority=100, *args, **kwargs): F9_AutoPart.__init__(self, writePriority=writePriority, *args, **kwargs) self.escrowcert = kwargs.get("escrowcert", "") self.backuppassphrase = kwargs.get("backuppassphrase", False) def __str__(self): retval = F9_AutoPart.__str__(self) if self.encrypted and self.escrowcert != "": retval = retval.strip() retval += " --escrowcert=\"%s\"" % self.escrowcert if self.backuppassphrase: retval += " --backuppassphrase" retval += "\n" return retval def _getParser(self): op = F9_AutoPart._getParser(self) op.add_option("--escrowcert") op.add_option("--backuppassphrase", action="store_true", default=False) return op ########NEW FILE######## __FILENAME__ = autostep # # Chris Lumens <clumens@redhat.com> # # Copyright 2005, 2006, 2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.options import * class FC3_AutoStep(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.op = self._getParser() self.autostep = kwargs.get("autostep", False) self.autoscreenshot = kwargs.get("autoscreenshot", False) def __str__(self): retval = KickstartCommand.__str__(self) if self.autostep: if self.autoscreenshot: retval += "autostep --autoscreenshot\n" else: retval += "autostep\n" return retval def _getParser(self): op = KSOptionParser() op.add_option("--autoscreenshot", dest="autoscreenshot", action="store_true", default=False) return op def parse(self, args): (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) self._setToSelf(self.op, opts) self.autostep = True return self ########NEW FILE######## __FILENAME__ = bootloader # # Chris Lumens <clumens@redhat.com> # # Copyright 2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.options import * class FC3_Bootloader(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=10, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.op = self._getParser() self.driveorder = kwargs.get("driveorder", []) self.appendLine = kwargs.get("appendLine", "") self.forceLBA = kwargs.get("forceLBA", False) self.linear = kwargs.get("linear", True) self.location = kwargs.get("location", "") self.md5pass = kwargs.get("md5pass", "") self.password = kwargs.get("password", "") self.upgrade = kwargs.get("upgrade", False) self.useLilo = kwargs.get("useLilo", False) self.deleteRemovedAttrs() def _getArgsAsStr(self): retval = "" if self.appendLine != "": retval += " --append=\"%s\"" % self.appendLine if self.linear: retval += " --linear" if self.location: retval += " --location=%s" % self.location if hasattr(self, "forceLBA") and self.forceLBA: retval += " --lba32" if self.password != "": retval += " --password=\"%s\"" % self.password if self.md5pass != "": retval += " --md5pass=\"%s\"" % self.md5pass if self.upgrade: retval += " --upgrade" if self.useLilo: retval += " --useLilo" if len(self.driveorder) > 0: retval += " --driveorder=\"%s\"" % ",".join(self.driveorder) return retval def __str__(self): retval = KickstartCommand.__str__(self) if self.location != "": retval += "# System bootloader configuration\nbootloader" retval += self._getArgsAsStr() + "\n" return retval def _getParser(self): def driveorder_cb (option, opt_str, value, parser): for d in value.split(','): parser.values.ensure_value(option.dest, []).append(d) op = KSOptionParser() op.add_option("--append", dest="appendLine") op.add_option("--linear", dest="linear", action="store_true", default=True) op.add_option("--nolinear", dest="linear", action="store_false") op.add_option("--location", dest="location", type="choice", default="mbr", choices=["mbr", "partition", "none", "boot"]) op.add_option("--lba32", dest="forceLBA", action="store_true", default=False) op.add_option("--password", dest="password", default="") op.add_option("--md5pass", dest="md5pass", default="") op.add_option("--upgrade", dest="upgrade", action="store_true", default=False) op.add_option("--useLilo", dest="useLilo", action="store_true", default=False) op.add_option("--driveorder", dest="driveorder", action="callback", callback=driveorder_cb, nargs=1, type="string") return op def parse(self, args): (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) self._setToSelf(self.op, opts) if self.currentCmd == "lilo": self.useLilo = True return self class FC4_Bootloader(FC3_Bootloader): removedKeywords = FC3_Bootloader.removedKeywords + ["linear", "useLilo"] removedAttrs = FC3_Bootloader.removedAttrs + ["linear", "useLilo"] def __init__(self, writePriority=10, *args, **kwargs): FC3_Bootloader.__init__(self, writePriority, *args, **kwargs) def _getArgsAsStr(self): retval = "" if self.appendLine != "": retval += " --append=\"%s\"" % self.appendLine if self.location: retval += " --location=%s" % self.location if hasattr(self, "forceLBA") and self.forceLBA: retval += " --lba32" if self.password != "": retval += " --password=\"%s\"" % self.password if self.md5pass != "": retval += " --md5pass=\"%s\"" % self.md5pass if self.upgrade: retval += " --upgrade" if len(self.driveorder) > 0: retval += " --driveorder=\"%s\"" % ",".join(self.driveorder) return retval def _getParser(self): op = FC3_Bootloader._getParser(self) op.remove_option("--linear") op.remove_option("--nolinear") op.remove_option("--useLilo") return op def parse(self, args): (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) self._setToSelf(self.op, opts) return self class F8_Bootloader(FC4_Bootloader): removedKeywords = FC4_Bootloader.removedKeywords removedAttrs = FC4_Bootloader.removedAttrs def __init__(self, writePriority=10, *args, **kwargs): FC4_Bootloader.__init__(self, writePriority, *args, **kwargs) self.timeout = kwargs.get("timeout", None) self.default = kwargs.get("default", "") def _getArgsAsStr(self): ret = FC4_Bootloader._getArgsAsStr(self) if self.timeout is not None: ret += " --timeout=%d" %(self.timeout,) if self.default: ret += " --default=%s" %(self.default,) return ret def _getParser(self): op = FC4_Bootloader._getParser(self) op.add_option("--timeout", dest="timeout", type="int") op.add_option("--default", dest="default") return op class F12_Bootloader(F8_Bootloader): removedKeywords = F8_Bootloader.removedKeywords removedAttrs = F8_Bootloader.removedAttrs def _getParser(self): op = F8_Bootloader._getParser(self) op.add_option("--lba32", dest="forceLBA", deprecated=1, action="store_true") return op class F14_Bootloader(F12_Bootloader): removedKeywords = F12_Bootloader.removedKeywords + ["forceLBA"] removedAttrs = F12_Bootloader.removedKeywords + ["forceLBA"] def _getParser(self): op = F12_Bootloader._getParser(self) op.remove_option("--lba32") return op class F15_Bootloader(F14_Bootloader): removedKeywords = F14_Bootloader.removedKeywords removedAttrs = F14_Bootloader.removedAttrs def __init__(self, writePriority=10, *args, **kwargs): F14_Bootloader.__init__(self, writePriority, *args, **kwargs) self.isCrypted = kwargs.get("isCrypted", False) def _getArgsAsStr(self): ret = F14_Bootloader._getArgsAsStr(self) if self.isCrypted: ret += " --iscrypted" return ret def _getParser(self): def password_cb(option, opt_str, value, parser): parser.values.isCrypted = True parser.values.password = value op = F14_Bootloader._getParser(self) op.add_option("--iscrypted", dest="isCrypted", action="store_true", default=False) op.add_option("--md5pass", action="callback", callback=password_cb, nargs=1, type="string") return op class RHEL5_Bootloader(FC4_Bootloader): removedKeywords = FC4_Bootloader.removedKeywords removedAttrs = FC4_Bootloader.removedAttrs def __init__(self, writePriority=10, *args, **kwargs): FC4_Bootloader.__init__(self, writePriority, *args, **kwargs) self.hvArgs = kwargs.get("hvArgs", "") def _getArgsAsStr(self): ret = FC4_Bootloader._getArgsAsStr(self) if self.hvArgs: ret += " --hvargs=\"%s\"" %(self.hvArgs,) return ret def _getParser(self): op = FC4_Bootloader._getParser(self) op.add_option("--hvargs", dest="hvArgs", type="string") return op class RHEL6_Bootloader(F12_Bootloader): removedKeywords = F12_Bootloader.removedKeywords removedAttrs = F12_Bootloader.removedAttrs def __init__(self, writePriority=10, *args, **kwargs): F12_Bootloader.__init__(self, writePriority, *args, **kwargs) self.isCrypted = kwargs.get("isCrypted", False) def _getArgsAsStr(self): ret = F12_Bootloader._getArgsAsStr(self) if self.isCrypted: ret += " --iscrypted" return ret def _getParser(self): def password_cb(option, opt_str, value, parser): parser.values.isCrypted = True parser.values.password = value op = F12_Bootloader._getParser(self) op.add_option("--iscrypted", dest="isCrypted", action="store_true", default=False) op.add_option("--md5pass", action="callback", callback=password_cb, nargs=1, type="string") return op ########NEW FILE######## __FILENAME__ = clearpart # # Chris Lumens <clumens@redhat.com> # # Copyright 2005, 2006, 2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.constants import * from pykickstart.errors import * from pykickstart.options import * class FC3_ClearPart(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=120, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.op = self._getParser() self.drives = kwargs.get("drives", []) self.initAll = kwargs.get("initAll", False) self.type = kwargs.get("type", None) def __str__(self): retval = KickstartCommand.__str__(self) if self.type is None: return retval if self.type == CLEARPART_TYPE_NONE: clearstr = "--none" elif self.type == CLEARPART_TYPE_LINUX: clearstr = "--linux" elif self.type == CLEARPART_TYPE_ALL: clearstr = "--all" else: clearstr = "" if self.initAll: initstr = "--initlabel" else: initstr = "" if len(self.drives) > 0: drivestr = "--drives=" + ",".join(self.drives) else: drivestr = "" retval += "# Partition clearing information\nclearpart %s %s %s\n" % (clearstr, initstr, drivestr) return retval def _getParser(self): def drive_cb (option, opt_str, value, parser): for d in value.split(','): parser.values.ensure_value(option.dest, []).append(d) op = KSOptionParser() op.add_option("--all", dest="type", action="store_const", const=CLEARPART_TYPE_ALL) op.add_option("--drives", dest="drives", action="callback", callback=drive_cb, nargs=1, type="string") op.add_option("--initlabel", dest="initAll", action="store_true", default=False) op.add_option("--linux", dest="type", action="store_const", const=CLEARPART_TYPE_LINUX) op.add_option("--none", dest="type", action="store_const", const=CLEARPART_TYPE_NONE) return op def parse(self, args): (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) self._setToSelf(self.op, opts) return self ########NEW FILE######## __FILENAME__ = device # # Chris Lumens <clumens@redhat.com> # # Copyright 2005, 2006, 2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.options import * import gettext import warnings _ = lambda x: gettext.ldgettext("pykickstart", x) class F8_DeviceData(BaseData): removedKeywords = BaseData.removedKeywords removedAttrs = BaseData.removedAttrs def __init__(self, *args, **kwargs): BaseData.__init__(self, *args, **kwargs) self.moduleName = kwargs.get("moduleName", "") self.moduleOpts = kwargs.get("moduleOpts", "") def __eq__(self, y): return self.moduleName == y.moduleName def __str__(self): retval = BaseData.__str__(self) if self.moduleName != "": retval += "device %s" % self.moduleName if self.moduleOpts != "": retval += " --opts=\"%s\"" % self.moduleOpts return retval + "\n" class FC3_Device(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.op = self._getParser() self.type = kwargs.get("type", "") self.moduleName = kwargs.get("moduleName", "") self.moduleOpts = kwargs.get("moduleOpts", "") def __eq__(self, y): return self.moduleName == y.moduleName def __str__(self): retval = KickstartCommand.__str__(self) if self.moduleName != "": retval += "device %s %s" % (self.type, self.moduleName) if self.moduleOpts != "": retval += " --opts=\"%s\"" % self.moduleOpts return retval + "\n" def _getParser(self): op = KSOptionParser() op.add_option("--opts", dest="moduleOpts", default="") return op def parse(self, args): (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) if len(extra) != 2: raise KickstartValueError, formatErrorMsg(self.lineno, msg=_("device command requires two arguments: module type and name")) self.moduleOpts = opts.moduleOpts self.type = extra[0] self.moduleName = extra[1] return self class F8_Device(FC3_Device): removedKeywords = FC3_Device.removedKeywords removedAttrs = FC3_Device.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): FC3_Device.__init__(self, writePriority, *args, **kwargs) self.deviceList = kwargs.get("deviceList", []) def __str__(self): retval = "" for device in self.deviceList: retval += device.__str__() return retval def parse(self, args): (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) if len(extra) != 1: raise KickstartValueError, formatErrorMsg(self.lineno, msg=_("%s command requires a single argument: %s") % ("device", "module name")) dd = F8_DeviceData() self._setToObj(self.op, opts, dd) dd.lineno = self.lineno dd.moduleName = extra[0] # Check for duplicates in the data list. if dd in self.dataList(): warnings.warn(_("A module with the name %s has already been defined.") % dd.moduleName) return dd def dataList(self): return self.deviceList ########NEW FILE######## __FILENAME__ = deviceprobe # # Chris Lumens <clumens@redhat.com> # # Copyright 2005, 2006, 2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * class FC3_DeviceProbe(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.deviceprobe = kwargs.get("deviceprobe", "") def __str__(self): retval = KickstartCommand.__str__(self) if self.deviceprobe != "": retval += "deviceprobe %s\n" % self.deviceprobe return retval def parse(self, args): self.deviceprobe = " ".join(args) return self ########NEW FILE######## __FILENAME__ = displaymode # # Chris Lumens <clumens@redhat.com> # # Copyright 2005, 2006, 2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.constants import * from pykickstart.options import * import gettext _ = lambda x: gettext.ldgettext("pykickstart", x) class FC3_DisplayMode(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.op = self._getParser() self.displayMode = kwargs.get("displayMode", None) def __str__(self): retval = KickstartCommand.__str__(self) if self.displayMode is None: return retval if self.displayMode == DISPLAY_MODE_CMDLINE: retval += "cmdline\n" elif self.displayMode == DISPLAY_MODE_GRAPHICAL: retval += "# Use graphical install\ngraphical\n" elif self.displayMode == DISPLAY_MODE_TEXT: retval += "# Use text mode install\ntext\n" return retval def _getParser(self): op = KSOptionParser() return op def parse(self, args): (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) if len(extra) > 0: raise KickstartParseError, formatErrorMsg(self.lineno, msg=_("Kickstart command %s does not take any arguments") % self.currentCmd) if self.currentCmd == "cmdline": self.displayMode = DISPLAY_MODE_CMDLINE elif self.currentCmd == "graphical": self.displayMode = DISPLAY_MODE_GRAPHICAL elif self.currentCmd == "text": self.displayMode = DISPLAY_MODE_TEXT return self ########NEW FILE######## __FILENAME__ = dmraid # # Chris Lumens <clumens@redhat.com> # Peter Jones <pjones@redhat.com> # # Copyright 2006, 2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.errors import * from pykickstart.options import * import gettext import warnings _ = lambda x: gettext.ldgettext("pykickstart", x) class FC6_DmRaidData(BaseData): removedKeywords = BaseData.removedKeywords removedAttrs = BaseData.removedAttrs def __init__(self, *args, **kwargs): BaseData.__init__(self, *args, **kwargs) self.name = kwargs.get("name", "") self.devices = kwargs.get("devices", []) self.dmset = kwargs.get("dmset", None) def __eq__(self, y): return self.name == y.name and self.devices == y.devices def __str__(self): retval = BaseData.__str__(self) retval += "dmraid --name=%s" % self.name for dev in self.devices: retval += " --dev=\"%s\"" % dev return retval + "\n" class FC6_DmRaid(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=60, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.op = self._getParser() self.dmraids = kwargs.get("dmraids", []) def __str__(self): retval = "" for dm in self.dmraids: retval += dm.__str__() return retval def _getParser(self): op = KSOptionParser() op.add_option("--name", dest="name", action="store", type="string", required=1) op.add_option("--dev", dest="devices", action="append", type="string", required=1) return op def parse(self, args): dm = FC6_DmRaidData() (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) dm.name = dm.name.split('/')[-1] self._setToObj(self.op, opts, dm) dm.lineno = self.lineno # Check for duplicates in the data list. if dm in self.dataList(): warnings.warn(_("A DM RAID device with the name %s and devices %s has already been defined.") % (dm.name, dm.devices)) return dm def dataList(self): return self.dmraids ########NEW FILE######## __FILENAME__ = driverdisk # # Chris Lumens <clumens@redhat.com> # # Copyright 2005, 2006, 2007, 2008 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.options import * import gettext _ = lambda x: gettext.ldgettext("pykickstart", x) class FC3_DriverDiskData(BaseData): removedKeywords = BaseData.removedKeywords removedAttrs = BaseData.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): BaseData.__init__(self, *args, **kwargs) self.partition = kwargs.get("partition", "") self.source = kwargs.get("source", "") self.type = kwargs.get("type", "") def _getArgsAsStr(self): retval = "" if self.partition: retval += "%s" % self.partition if hasattr(self, "type") and self.type: retval += " --type=%s" % self.type elif self.source: retval += "--source=%s" % self.source return retval def __str__(self): retval = BaseData.__str__(self) retval += "driverdisk %s\n" % self._getArgsAsStr() return retval class FC4_DriverDiskData(FC3_DriverDiskData): removedKeywords = FC3_DriverDiskData.removedKeywords removedAttrs = FC3_DriverDiskData.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): FC3_DriverDiskData.__init__(self, *args, **kwargs) self.deleteRemovedAttrs() self.biospart = kwargs.get("biospart", "") def _getArgsAsStr(self): retval = "" if self.partition: retval += "%s" % self.partition if hasattr(self, "type") and self.type: retval += " --type=%s" % self.type elif self.source: retval += "--source=%s" % self.source elif self.biospart: retval += "--biospart=%s" % self.biospart return retval class F12_DriverDiskData(FC4_DriverDiskData): removedKeywords = FC4_DriverDiskData.removedKeywords + ["type"] removedAttrs = FC4_DriverDiskData.removedAttrs + ["type"] def __init__(self, *args, **kwargs): FC4_DriverDiskData.__init__(self, *args, **kwargs) self.deleteRemovedAttrs() F14_DriverDiskData = F12_DriverDiskData class FC3_DriverDisk(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.op = self._getParser() self.driverdiskList = kwargs.get("driverdiskList", []) def __str__(self): retval = "" for dd in self.driverdiskList: retval += dd.__str__() return retval def _getParser(self): op = KSOptionParser() op.add_option("--source") op.add_option("--type") return op def parse(self, args): (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) if len(extra) > 1: raise KickstartValueError, formatErrorMsg(self.lineno, msg=_("Only one partition may be specified for driverdisk command.")) if len(extra) == 1 and opts.source: raise KickstartValueError, formatErrorMsg(self.lineno, msg=_("Only one of --source and partition may be specified for driverdisk command.")) if not extra and not opts.source: raise KickstartValueError, formatErrorMsg(self.lineno, msg=_("One of --source or partition must be specified for driverdisk command.")) ddd = self.handler.DriverDiskData() self._setToObj(self.op, opts, ddd) ddd.lineno = self.lineno if len(extra) == 1: ddd.partition = extra[0] return ddd def dataList(self): return self.driverdiskList class FC4_DriverDisk(FC3_DriverDisk): removedKeywords = FC3_DriverDisk.removedKeywords removedAttrs = FC3_DriverDisk.removedKeywords def _getParser(self): op = FC3_DriverDisk._getParser(self) op.add_option("--biospart") return op def parse(self, args): (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) if len(extra) > 1: raise KickstartValueError, formatErrorMsg(self.lineno, msg=_("Only one partition may be specified for driverdisk command.")) if len(extra) == 1 and opts.source: raise KickstartValueError, formatErrorMsg(self.lineno, msg=_("Only one of --source and partition may be specified for driverdisk command.")) elif len(extra) == 1 and opts.biospart: raise KickstartValueError, formatErrorMsg(self.lineno, msg=_("Only one of --biospart and partition may be specified for driverdisk command.")) elif opts.source and opts.biospart: raise KickstartValueError, formatErrorMsg(self.lineno, msg=_("Only one of --biospart and --source may be specified for driverdisk command.")) if not extra and not opts.source and not opts.biospart: raise KickstartValueError, formatErrorMsg(self.lineno, msg=_("One of --source, --biospart, or partition must be specified for driverdisk command.")) ddd = self.handler.DriverDiskData() self._setToObj(self.op, opts, ddd) ddd.lineno = self.lineno if len(extra) == 1: ddd.partition = extra[0] return ddd class F12_DriverDisk(FC4_DriverDisk): removedKeywords = FC4_DriverDisk.removedKeywords removedAttrs = FC4_DriverDisk.removedKeywords def _getParser(self): op = FC4_DriverDisk._getParser(self) op.add_option("--type", deprecated=1) return op class F14_DriverDisk(F12_DriverDisk): removedKeywords = F12_DriverDisk.removedKeywords removedAttrs = F12_DriverDisk.removedKeywords def _getParser(self): op = F12_DriverDisk._getParser(self) op.remove_option("--type") return op ########NEW FILE######## __FILENAME__ = fcoe # # Hans de Goede <hdegoede@redhat.com> # # Copyright 2009 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.options import * import gettext import warnings _ = lambda x: gettext.ldgettext("pykickstart", x) class F12_FcoeData(BaseData): removedKeywords = BaseData.removedKeywords removedAttrs = BaseData.removedAttrs def __init__(self, *args, **kwargs): BaseData.__init__(self, *args, **kwargs) self.nic = kwargs.get("nic", None) def __eq__(self, y): return self.nic == y.nic def _getArgsAsStr(self): retval = "" if self.nic: retval += " --nic=%s" % self.nic return retval def __str__(self): retval = BaseData.__str__(self) retval += "fcoe%s\n" % self._getArgsAsStr() return retval class F13_FcoeData(F12_FcoeData): removedKeywords = F12_FcoeData.removedKeywords removedAttrs = F12_FcoeData.removedAttrs def __init__(self, *args, **kwargs): F12_FcoeData.__init__(self, *args, **kwargs) self.dcb = kwargs.get("dcb", False) def _getArgsAsStr(self): retval = F12_FcoeData._getArgsAsStr(self) if self.dcb: retval += " --dcb" return retval class F12_Fcoe(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=71, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.op = self._getParser() self.fcoe = kwargs.get("fcoe", []) def __str__(self): retval = "" for fcoe in self.fcoe: retval += fcoe.__str__() return retval def _getParser(self): op = KSOptionParser() op.add_option("--nic", dest="nic", required=1) return op def parse(self, args): zd = self.handler.FcoeData() (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) if len(extra) > 0: mapping = {"command": "fcoe", "options": extra} raise KickstartValueError, formatErrorMsg(self.lineno, msg=_("Unexpected arguments to %(command)s command: %(options)s") % mapping) self._setToObj(self.op, opts, zd) zd.lineno = self.lineno # Check for duplicates in the data list. if zd in self.dataList(): warnings.warn(_("A FCOE device with the name %s has already been defined.") % zd.nic) return zd def dataList(self): return self.fcoe class F13_Fcoe(F12_Fcoe): removedKeywords = F12_Fcoe.removedKeywords removedAttrs = F12_Fcoe.removedAttrs def _getParser(self): op = F12_Fcoe._getParser(self) op.add_option("--dcb", dest="dcb", action="store_true", default=False) return op ########NEW FILE######## __FILENAME__ = firewall # # Chris Lumens <clumens@redhat.com> # # Copyright 2005, 2006, 2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.errors import * from pykickstart.options import * import gettext _ = lambda x: gettext.ldgettext("pykickstart", x) class FC3_Firewall(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.op = self._getParser() self.enabled = kwargs.get("enabled", None) self.ports = kwargs.get("ports", []) self.trusts = kwargs.get("trusts", []) def __str__(self): extra = [] filteredPorts = [] retval = KickstartCommand.__str__(self) if self.enabled is None: return retval if self.enabled: # It's possible we have words in the ports list instead of # port:proto (s-c-kickstart may do this). So, filter those # out into their own list leaving what we expect. for port in self.ports: if port == "ssh": extra.append(" --ssh") elif port == "telnet": extra.append(" --telnet") elif port == "smtp": extra.append(" --smtp") elif port == "http": extra.append(" --http") elif port == "ftp": extra.append(" --ftp") else: filteredPorts.append(port) # All the port:proto strings go into a comma-separated list. portstr = ",".join(filteredPorts) if len(portstr) > 0: portstr = " --port=" + portstr else: portstr = "" extrastr = "".join(extra) truststr = ",".join(self.trusts) if len(truststr) > 0: truststr = " --trust=" + truststr # The output port list consists only of port:proto for # everything that we don't recognize, and special options for # those that we do. retval += "# Firewall configuration\nfirewall --enabled%s%s%s\n" % (extrastr, portstr, truststr) else: retval += "# Firewall configuration\nfirewall --disabled\n" return retval def _getParser(self): def firewall_port_cb (option, opt_str, value, parser): for p in value.split(","): p = p.strip() if p.find(":") == -1: p = "%s:tcp" % p parser.values.ensure_value(option.dest, []).append(p) op = KSOptionParser(mapping={"ssh":["22:tcp"], "telnet":["23:tcp"], "smtp":["25:tcp"], "http":["80:tcp", "443:tcp"], "ftp":["21:tcp"]}) op.add_option("--disable", "--disabled", dest="enabled", action="store_false") op.add_option("--enable", "--enabled", dest="enabled", action="store_true", default=True) op.add_option("--ftp", "--http", "--smtp", "--ssh", "--telnet", dest="ports", action="map_extend") op.add_option("--high", deprecated=1) op.add_option("--medium", deprecated=1) op.add_option("--port", dest="ports", action="callback", callback=firewall_port_cb, nargs=1, type="string") op.add_option("--trust", dest="trusts", action="append") return op def parse(self, args): (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) if len(extra) != 0: mapping = {"command": "firewall", "options": extra} raise KickstartValueError, formatErrorMsg(self.lineno, msg=_("Unexpected arguments to %(command)s command: %(options)s") % mapping) self._setToSelf(self.op, opts) return self class F9_Firewall(FC3_Firewall): removedKeywords = FC3_Firewall.removedKeywords removedAttrs = FC3_Firewall.removedAttrs def _getParser(self): op = FC3_Firewall._getParser(self) op.remove_option("--high") op.remove_option("--medium") return op class F10_Firewall(F9_Firewall): removedKeywords = F9_Firewall.removedKeywords removedAttrs = F9_Firewall.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): F9_Firewall.__init__(self, writePriority, *args, **kwargs) self.services = kwargs.get("services", []) def __str__(self): if self.enabled is None: return "" retval = F9_Firewall.__str__(self) if self.enabled: retval = retval.strip() svcstr = ",".join(self.services) if len(svcstr) > 0: svcstr = " --service=" + svcstr else: svcstr = "" return retval + "%s\n" % svcstr else: return retval def _getParser(self): def service_cb (option, opt_str, value, parser): # python2.4 does not support action="append_const" that we were # using for these options. Instead, we have to fake it by # appending whatever the option string is to the service list. if not value: parser.values.ensure_value(option.dest, []).append(opt_str[2:]) return for p in value.split(","): p = p.strip() parser.values.ensure_value(option.dest, []).append(p) op = F9_Firewall._getParser(self) op.add_option("--service", dest="services", action="callback", callback=service_cb, nargs=1, type="string") op.add_option("--ftp", dest="services", action="callback", callback=service_cb) op.add_option("--http", dest="services", action="callback", callback=service_cb) op.add_option("--smtp", dest="services", action="callback", callback=service_cb) op.add_option("--ssh", dest="services", action="callback", callback=service_cb) op.add_option("--telnet", deprecated=1) return op class F14_Firewall(F10_Firewall): removedKeywords = F10_Firewall.removedKeywords + ["telnet"] removedAttrs = F10_Firewall.removedAttrs + ["telnet"] def _getParser(self): op = F10_Firewall._getParser(self) op.remove_option("--telnet") return op ########NEW FILE######## __FILENAME__ = firstboot # # Chris Lumens <clumens@redhat.com> # # Copyright 2005, 2006, 2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.constants import * from pykickstart.options import * class FC3_Firstboot(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.op = self._getParser() self.firstboot = kwargs.get("firstboot", None) def __str__(self): retval = KickstartCommand.__str__(self) if self.firstboot is None: return retval if self.firstboot == FIRSTBOOT_SKIP: retval += "firstboot --disable\n" elif self.firstboot == FIRSTBOOT_DEFAULT: retval += "# Run the Setup Agent on first boot\nfirstboot --enable\n" elif self.firstboot == FIRSTBOOT_RECONFIG: retval += "# Run the Setup Agent on first boot\nfirstboot --reconfig\n" return retval def _getParser(self): op = KSOptionParser() op.add_option("--disable", "--disabled", dest="firstboot", action="store_const", const=FIRSTBOOT_SKIP) op.add_option("--enable", "--enabled", dest="firstboot", action="store_const", const=FIRSTBOOT_DEFAULT) op.add_option("--reconfig", dest="firstboot", action="store_const", const=FIRSTBOOT_RECONFIG) return op def parse(self, args): (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) self.firstboot = opts.firstboot return self ########NEW FILE######## __FILENAME__ = group # # Chris Lumens <clumens@redhat.com> # # Copyright 2009 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.constants import * from pykickstart.errors import * from pykickstart.options import * import gettext import warnings _ = lambda x: gettext.ldgettext("pykickstart", x) class F12_GroupData(BaseData): removedKeywords = BaseData.removedKeywords removedAttrs = BaseData.removedAttrs def __init__(self, *args, **kwargs): BaseData.__init__(self, *args, **kwargs) self.name = kwargs.get("name", "") self.gid = kwargs.get("gid", None) def __eq__(self, y): return self.name == y.name def __str__(self): retval = BaseData.__str__(self) retval += "group" if self.name: retval += " --name=%s" % self.name if self.gid: retval += " --gid=%s" % self.gid return retval + "\n" class F12_Group(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.op = self._getParser() self.groupList = kwargs.get("groupList", []) def __str__(self): retval = "" for user in self.groupList: retval += user.__str__() return retval def _getParser(self): op = KSOptionParser() op.add_option("--name", required=1) op.add_option("--gid", type="int") return op def parse(self, args): gd = self.handler.GroupData() (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) self._setToObj(self.op, opts, gd) gd.lineno = self.lineno # Check for duplicates in the data list. if gd in self.dataList(): warnings.warn(_("A group with the name %s has already been defined.") % gd.name) return gd def dataList(self): return self.groupList ########NEW FILE######## __FILENAME__ = ignoredisk # # Chris Lumens <clumens@redhat.com> # # Copyright 2005, 2006, 2007, 2009 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.options import * import gettext _ = lambda x: gettext.ldgettext("pykickstart", x) class FC3_IgnoreDisk(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.op = self._getParser() self.ignoredisk = kwargs.get("ignoredisk", []) def __str__(self): retval = KickstartCommand.__str__(self) if len(self.ignoredisk) > 0: retval += "ignoredisk --drives=%s\n" % ",".join(self.ignoredisk) return retval def _getParser(self): def drive_cb (option, opt_str, value, parser): for d in value.split(','): parser.values.ensure_value(option.dest, []).append(d) op = KSOptionParser() op.add_option("--drives", dest="ignoredisk", action="callback", callback=drive_cb, nargs=1, type="string", required=1) return op def parse(self, args): (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) self._setToSelf(self.op, opts) return self class F8_IgnoreDisk(FC3_IgnoreDisk): removedKeywords = FC3_IgnoreDisk.removedKeywords removedAttrs = FC3_IgnoreDisk.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): FC3_IgnoreDisk.__init__(self, writePriority, *args, **kwargs) self.onlyuse = kwargs.get("onlyuse", []) def __str__(self): retval = KickstartCommand.__str__(self) if len(self.ignoredisk) > 0: retval += "ignoredisk --drives=%s\n" % ",".join(self.ignoredisk) elif len(self.onlyuse) > 0: retval += "ignoredisk --only-use=%s\n" % ",".join(self.onlyuse) return retval def parse(self, args, errorCheck=True): retval = FC3_IgnoreDisk.parse(self, args) if errorCheck: if (len(self.ignoredisk) == 0 and len(self.onlyuse) == 0) or (len(self.ignoredisk) > 0 and (len(self.onlyuse) > 0)): raise KickstartValueError, formatErrorMsg(self.lineno, msg=_("One of --drives or --only-use must be specified for ignoredisk command.")) return retval def _getParser(self): def drive_cb (option, opt_str, value, parser): for d in value.split(','): parser.values.ensure_value(option.dest, []).append(d) op = FC3_IgnoreDisk._getParser(self) op.add_option("--drives", dest="ignoredisk", action="callback", callback=drive_cb, nargs=1, type="string") op.add_option("--only-use", dest="onlyuse", action="callback", callback=drive_cb, nargs=1, type="string") return op class RHEL6_IgnoreDisk(F8_IgnoreDisk): removedKeywords = F8_IgnoreDisk.removedKeywords removedAttrs = F8_IgnoreDisk.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): F8_IgnoreDisk.__init__(self, writePriority, *args, **kwargs) self.interactive = kwargs.get("interactive", False) if self.interactive: self.ignoredisk = [] def __str__(self): retval = F8_IgnoreDisk.__str__(self) if self.interactive: retval = "ignoredisk --interactive\n" return retval def parse(self, args): retval = F8_IgnoreDisk.parse(self, args, errorCheck=False) howmany = 0 if len(self.ignoredisk) > 0: howmany += 1 if len(self.onlyuse) > 0: howmany += 1 if self.interactive: howmany += 1 if howmany != 1: raise KickstartValueError, formatErrorMsg(self.lineno, msg=_("One of --drives , --only-use , or --interactive must be specified for ignoredisk command.")) return retval def _getParser(self): op = F8_IgnoreDisk._getParser(self) op.add_option("--interactive", dest="interactive", action="store_true", default=False) return op F14_IgnoreDisk = RHEL6_IgnoreDisk ########NEW FILE######## __FILENAME__ = interactive # # Chris Lumens <clumens@redhat.com> # # Copyright 2005, 2006, 2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.errors import * from pykickstart.options import * import gettext _ = lambda x: gettext.ldgettext("pykickstart", x) class FC3_Interactive(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.op = self._getParser() self.interactive = kwargs.get("interactive", False) def __str__(self): retval = KickstartCommand.__str__(self) if self.interactive: retval += "# Use interactive kickstart installation method\ninteractive\n" return retval def _getParser(self): op = KSOptionParser() return op def parse(self, args): (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) if len(extra) > 0: raise KickstartValueError, formatErrorMsg(self.lineno, msg=_("Kickstart command %s does not take any arguments") % "interactive") self.interactive = True return self class F14_Interactive(DeprecatedCommand): def __init__(self): DeprecatedCommand.__init__(self) ########NEW FILE######## __FILENAME__ = iscsi # # Chris Lumens <clumens@redhat.com> # Peter Jones <pjones@redhat.com> # # Copyright 2005, 2006, 2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.errors import * from pykickstart.options import * import gettext _ = lambda x: gettext.ldgettext("pykickstart", x) class FC6_IscsiData(BaseData): removedKeywords = BaseData.removedKeywords removedAttrs = BaseData.removedAttrs def __init__(self, *args, **kwargs): BaseData.__init__(self, *args, **kwargs) self.ipaddr = kwargs.get("ipaddr", "") self.port = kwargs.get("port", "3260") self.target = kwargs.get("target", "") self.user = kwargs.get("user", None) self.password = kwargs.get("password", None) def _getArgsAsStr(self): retval = "" if self.target != "": retval += " --target=%s" % self.target if self.ipaddr != "": retval += " --ipaddr=%s" % self.ipaddr if self.port != "3260": retval += " --port=%s" % self.port if self.user is not None: retval += " --user=%s" % self.user if self.password is not None: retval += " --password=%s" % self.password return retval def __str__(self): retval = BaseData.__str__(self) retval += "iscsi%s\n" % self._getArgsAsStr() return retval class F10_IscsiData(FC6_IscsiData): removedKeywords = FC6_IscsiData.removedKeywords removedAttrs = FC6_IscsiData.removedAttrs def __init__(self, *args, **kwargs): FC6_IscsiData.__init__(self, *args, **kwargs) self.user_in = kwargs.get("user_in", None) self.password_in = kwargs.get("password_in", None) def _getArgsAsStr(self): retval = FC6_IscsiData._getArgsAsStr(self) if self.user_in is not None: retval += " --reverse-user=%s" % self.user_in if self.password_in is not None: retval += " --reverse-password=%s" % self.password_in return retval class FC6_Iscsi(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=71, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.op = self._getParser() self.iscsi = kwargs.get("iscsi", []) def __str__(self): retval = "" for iscsi in self.iscsi: retval += iscsi.__str__() return retval def _getParser(self): op = KSOptionParser() op.add_option("--target", dest="target", action="store", type="string") op.add_option("--ipaddr", dest="ipaddr", action="store", type="string", required=1) op.add_option("--port", dest="port", action="store", type="string") op.add_option("--user", dest="user", action="store", type="string") op.add_option("--password", dest="password", action="store", type="string") return op def parse(self, args): (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) if len(extra) != 0: mapping = {"command": "iscsi", "options": extra} raise KickstartValueError, formatErrorMsg(self.lineno, msg=_("Unexpected arguments to %(command)s command: %(options)s") % mapping) dd = self.handler.IscsiData() self._setToObj(self.op, opts, dd) dd.lineno = self.lineno return dd def dataList(self): return self.iscsi class F10_Iscsi(FC6_Iscsi): removedKeywords = FC6_Iscsi.removedKeywords removedAttrs = FC6_Iscsi.removedAttrs def _getParser(self): op = FC6_Iscsi._getParser(self) op.add_option("--reverse-user", dest="user_in", action="store", type="string") op.add_option("--reverse-password", dest="password_in", action="store", type="string") return op ########NEW FILE######## __FILENAME__ = iscsiname # # Chris Lumens <clumens@redhat.com> # Peter Jones <pjones@redhat.com> # # Copyright 2006, 2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.errors import * from pykickstart.options import * import gettext _ = lambda x: gettext.ldgettext("pykickstart", x) class FC6_IscsiName(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=70, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.op = self._getParser() self.iscsiname = kwargs.get("iscsiname", "") def __str__(self): retval = KickstartCommand.__str__(self) if self.iscsiname != "": retval += "iscsiname %s\n" % self.iscsiname return retval def _getParser(self): op = KSOptionParser() return op def parse(self, args): (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) if len(extra) != 1: raise KickstartValueError, formatErrorMsg(self.lineno, msg=_("Kickstart command %s requires one argument") % "iscsiname") self.iscsiname = extra[0] return self ########NEW FILE######## __FILENAME__ = key # # Chris Lumens <clumens@redhat.com> # # Copyright 2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.constants import * from pykickstart.errors import * from pykickstart.options import * import gettext _ = lambda x: gettext.ldgettext("pykickstart", x) class RHEL5_Key(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.op = self._getParser() self.key = kwargs.get("key", "") self.skip = kwargs.get("skip", False) def __str__(self): retval = KickstartCommand.__str__(self) if self.key == KS_INSTKEY_SKIP: retval += "key --skip\n" elif self.key != "": retval += "key %s\n" % self.key return retval def _getParser(self): op = KSOptionParser() op.add_option("--skip", action="store_true", default=False) return op def parse(self, args): (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) self._setToSelf(self.op, opts) if self.skip: self.key = KS_INSTKEY_SKIP elif len(extra) != 1: raise KickstartValueError, formatErrorMsg(self.lineno, msg=_("Kickstart command %s requires one argument") % "key") else: self.key = extra[0] return self ########NEW FILE######## __FILENAME__ = keyboard # # Chris Lumens <clumens@redhat.com> # # Copyright 2005, 2006, 2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.errors import * from pykickstart.options import * import gettext _ = lambda x: gettext.ldgettext("pykickstart", x) class FC3_Keyboard(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.op = self._getParser() self.keyboard = kwargs.get("keyboard", "") def __str__(self): retval = KickstartCommand.__str__(self) if self.keyboard != "": retval += "# System keyboard\nkeyboard %s\n" % self.keyboard return retval def _getParser(self): op = KSOptionParser() return op def parse(self, args): (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) if len(extra) != 1: raise KickstartValueError, formatErrorMsg(self.lineno, msg=_("Kickstart command %s requires one argument") % "keyboard") self.keyboard = extra[0] return self ########NEW FILE######## __FILENAME__ = lang # # Chris Lumens <clumens@redhat.com> # # Copyright 2005, 2006, 2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.errors import * from pykickstart.options import * import gettext _ = lambda x: gettext.ldgettext("pykickstart", x) class FC3_Lang(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.op = self._getParser() self.lang = kwargs.get("lang", "") def __str__(self): retval = KickstartCommand.__str__(self) if self.lang != "": retval += "# System language\nlang %s\n" % self.lang return retval def _getParser(self): op = KSOptionParser() return op def parse(self, args): (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) if len(extra) != 1: raise KickstartValueError, formatErrorMsg(self.lineno, msg=_("Kickstart command %s requires one argument") % "lang") self.lang = extra[0] return self def apply(self, instroot="/"): if self.lang == "": return f = open(instroot + "/etc/sysconfig/i18n", "w+") f.write("LANG=\"%s\"\n" %(self.lang,)) f.close() ########NEW FILE######## __FILENAME__ = langsupport # # Chris Lumens <clumens@redhat.com> # # Copyright 2005, 2006, 2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.options import * class FC3_LangSupport(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.op = self._getParser() self.deflang = kwargs.get("deflang", "") self.supported = kwargs.get("supported", []) def __str__(self): retval = KickstartCommand.__str__(self) if self.deflang: retval += "langsupport --default=%s" % self.deflang if self.supported: retval += " %s" % " ".join(self.supported) return retval + "\n" def _getParser(self): op = KSOptionParser() op.add_option("--default", dest="deflang", default="en_US.UTF-8") return op def parse(self, args): (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) self._setToSelf(self.op, opts) self.supported = extra return self class FC5_LangSupport(DeprecatedCommand): def __init__(self): DeprecatedCommand.__init__(self) ########NEW FILE######## __FILENAME__ = lilocheck # # Chris Lumens <clumens@redhat.com> # # Copyright 2005, 2006, 2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.errors import * from pykickstart.options import * import gettext _ = lambda x: gettext.ldgettext("pykickstart", x) class FC3_LiloCheck(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.op = self._getParser() self.check = kwargs.get("check", False) def __str__(self): retval = KickstartCommand.__str__(self) if self.check: retval += "lilocheck\n" return retval def _getParser(self): op = KSOptionParser() return op def parse(self, args): (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) if len(extra) > 0: raise KickstartValueError, formatErrorMsg(self.lineno, msg=_("Kickstart command %s does not take any arguments") % "lilocheck") self.check = True return self ########NEW FILE######## __FILENAME__ = logging # # Chris Lumens <clumens@redhat.com> # # Copyright 2007, 2009 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.errors import * from pykickstart.options import * import gettext _ = lambda x: gettext.ldgettext("pykickstart", x) class FC6_Logging(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.op = self._getParser() self.host = kwargs.get("host", "") self.level = kwargs.get("level", "info") self.port = kwargs.get("port", "") def __str__(self): retval = KickstartCommand.__str__(self) retval += "# Installation logging level\nlogging --level=%s" % self.level if self.host != "": retval += " --host=%s" % self.host if self.port != "": retval += " --port=%s" % self.port return retval + "\n" def _getParser(self): op = KSOptionParser() op.add_option("--host") op.add_option("--level", type="choice", default="info", choices=["debug", "info", "warning", "error", "critical"]) op.add_option("--port") return op def parse(self, args): (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) if opts.port and not opts.host: raise KickstartParseError, formatErrorMsg(self.lineno, msg=_("Can't specify --port without --host.")) self._setToSelf(self.op, opts) return self ########NEW FILE######## __FILENAME__ = logvol # # Chris Lumens <clumens@redhat.com> # # Copyright 2005, 2006, 2007, 2008 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.errors import * from pykickstart.options import * import gettext import warnings _ = lambda x: gettext.ldgettext("pykickstart", x) class FC3_LogVolData(BaseData): removedKeywords = BaseData.removedKeywords removedAttrs = BaseData.removedAttrs def __init__(self, *args, **kwargs): BaseData.__init__(self, *args, **kwargs) self.fstype = kwargs.get("fstype", "") self.grow = kwargs.get("grow", False) self.maxSizeMB = kwargs.get("maxSizeMB", 0) self.name = kwargs.get("name", "") self.format = kwargs.get("format", True) self.percent = kwargs.get("percent", 0) self.recommended = kwargs.get("recommended", False) self.size = kwargs.get("size", None) self.preexist = kwargs.get("preexist", False) self.vgname = kwargs.get("vgname", "") self.mountpoint = kwargs.get("mountpoint", "") def __eq__(self, y): return self.vgname == y.vgname and self.name == y.name def _getArgsAsStr(self): retval = "" if self.fstype != "": retval += " --fstype=\"%s\"" % self.fstype if self.grow: retval += " --grow" if self.maxSizeMB > 0: retval += " --maxsize=%d" % self.maxSizeMB if not self.format: retval += " --noformat" if self.percent > 0: retval += " --percent=%d" % self.percent if self.recommended: retval += " --recommended" if self.size > 0: retval += " --size=%d" % self.size if self.preexist: retval += " --useexisting" return retval def __str__(self): retval = BaseData.__str__(self) retval += "logvol %s %s --name=%s --vgname=%s\n" % (self.mountpoint, self._getArgsAsStr(), self.name, self.vgname) return retval class FC4_LogVolData(FC3_LogVolData): removedKeywords = FC3_LogVolData.removedKeywords removedAttrs = FC3_LogVolData.removedAttrs def __init__(self, *args, **kwargs): FC3_LogVolData.__init__(self, *args, **kwargs) self.bytesPerInode = kwargs.get("bytesPerInode", 4096) self.fsopts = kwargs.get("fsopts", "") def _getArgsAsStr(self): retval = FC3_LogVolData._getArgsAsStr(self) if hasattr(self, "bytesPerInode") and self.bytesPerInode != 0: retval += " --bytes-per-inode=%d" % self.bytesPerInode if self.fsopts != "": retval += " --fsoptions=\"%s\"" % self.fsopts return retval class RHEL5_LogVolData(FC4_LogVolData): removedKeywords = FC4_LogVolData.removedKeywords removedAttrs = FC4_LogVolData.removedAttrs def __init__(self, *args, **kwargs): FC4_LogVolData.__init__(self, *args, **kwargs) self.encrypted = kwargs.get("encrypted", False) self.passphrase = kwargs.get("passphrase", "") def _getArgsAsStr(self): retval = FC4_LogVolData._getArgsAsStr(self) if self.encrypted: retval += " --encrypted" if self.passphrase != "": retval += " --passphrase=\"%s\"" % self.passphrase return retval class F9_LogVolData(FC4_LogVolData): removedKeywords = FC4_LogVolData.removedKeywords + ["bytesPerInode"] removedAttrs = FC4_LogVolData.removedAttrs + ["bytesPerInode"] def __init__(self, *args, **kwargs): FC4_LogVolData.__init__(self, *args, **kwargs) self.deleteRemovedAttrs() self.fsopts = kwargs.get("fsopts", "") self.fsprofile = kwargs.get("fsprofile", "") self.encrypted = kwargs.get("encrypted", False) self.passphrase = kwargs.get("passphrase", "") def _getArgsAsStr(self): retval = FC4_LogVolData._getArgsAsStr(self) if self.fsprofile != "": retval += " --fsprofile=\"%s\"" % self.fsprofile if self.encrypted: retval += " --encrypted" if self.passphrase != "": retval += " --passphrase=\"%s\"" % self.passphrase return retval class F12_LogVolData(F9_LogVolData): removedKeywords = F9_LogVolData.removedKeywords removedAttrs = F9_LogVolData.removedAttrs def __init__(self, *args, **kwargs): F9_LogVolData.__init__(self, *args, **kwargs) self.deleteRemovedAttrs() self.escrowcert = kwargs.get("escrowcert", "") self.backuppassphrase = kwargs.get("backuppassphrase", False) def _getArgsAsStr(self): retval = F9_LogVolData._getArgsAsStr(self) if self.encrypted and self.escrowcert != "": retval += " --escrowcert=\"%s\"" % self.escrowcert if self.backuppassphrase: retval += " --backuppassphrase" return retval F14_LogVolData = F12_LogVolData class F15_LogVolData(F14_LogVolData): removedKeywords = F14_LogVolData.removedKeywords removedAttrs = F14_LogVolData.removedAttrs def __init__(self, *args, **kwargs): F14_LogVolData.__init__(self, *args, **kwargs) self.label = kwargs.get("label", "") def _getArgsAsStr(self): retval = F14_LogVolData._getArgsAsStr(self) if self.label != "": retval += " --label=\"%s\"" % self.label return retval class FC3_LogVol(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=133, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.op = self._getParser() self.lvList = kwargs.get("lvList", []) def __str__(self): retval = "" for part in self.lvList: retval += part.__str__() return retval def _getParser(self): def lv_cb (option, opt_str, value, parser): parser.values.format = False parser.values.preexist = True op = KSOptionParser() op.add_option("--fstype", dest="fstype") op.add_option("--grow", dest="grow", action="store_true", default=False) op.add_option("--maxsize", dest="maxSizeMB", action="store", type="int", nargs=1) op.add_option("--name", dest="name", required=1) op.add_option("--noformat", action="callback", callback=lv_cb, dest="format", default=True, nargs=0) op.add_option("--percent", dest="percent", action="store", type="int", nargs=1) op.add_option("--recommended", dest="recommended", action="store_true", default=False) op.add_option("--size", dest="size", action="store", type="int", nargs=1) op.add_option("--useexisting", dest="preexist", action="store_true", default=False) op.add_option("--vgname", dest="vgname", required=1) return op def parse(self, args): (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) if len(extra) == 0: raise KickstartValueError, formatErrorMsg(self.lineno, msg=_("Mount point required for %s") % "logvol") lvd = self.handler.LogVolData() self._setToObj(self.op, opts, lvd) lvd.lineno = self.lineno lvd.mountpoint=extra[0] # Check for duplicates in the data list. if lvd in self.dataList(): warnings.warn(_("A logical volume with the name %s has already been defined in volume group %s.") % (lvd.device, lvd.vgname)) return lvd def dataList(self): return self.lvList class FC4_LogVol(FC3_LogVol): removedKeywords = FC3_LogVol.removedKeywords removedAttrs = FC3_LogVol.removedAttrs def _getParser(self): op = FC3_LogVol._getParser(self) op.add_option("--bytes-per-inode", dest="bytesPerInode", action="store", type="int", nargs=1) op.add_option("--fsoptions", dest="fsopts") return op class RHEL5_LogVol(FC4_LogVol): removedKeywords = FC4_LogVol.removedKeywords removedAttrs = FC4_LogVol.removedAttrs def _getParser(self): op = FC4_LogVol._getParser(self) op.add_option("--encrypted", action="store_true", default=False) op.add_option("--passphrase") return op class F9_LogVol(FC4_LogVol): removedKeywords = FC4_LogVol.removedKeywords removedAttrs = FC4_LogVol.removedAttrs def _getParser(self): op = FC4_LogVol._getParser(self) op.add_option("--bytes-per-inode", deprecated=1) op.add_option("--fsprofile", dest="fsprofile", action="store", type="string", nargs=1) op.add_option("--encrypted", action="store_true", default=False) op.add_option("--passphrase") return op class F12_LogVol(F9_LogVol): removedKeywords = F9_LogVol.removedKeywords removedAttrs = F9_LogVol.removedAttrs def _getParser(self): op = F9_LogVol._getParser(self) op.add_option("--escrowcert") op.add_option("--backuppassphrase", action="store_true", default=False) return op class F14_LogVol(F12_LogVol): removedKeywords = F12_LogVol.removedKeywords removedAttrs = F12_LogVol.removedAttrs def _getParser(self): op = F12_LogVol._getParser(self) op.remove_option("--bytes-per-inode") return op class F15_LogVol(F14_LogVol): removedKeywords = F14_LogVol.removedKeywords removedAttrs = F14_LogVol.removedAttrs def _getParser(self): op = F14_LogVol._getParser(self) op.add_option("--label") return op ########NEW FILE######## __FILENAME__ = mediacheck # # Chris Lumens <clumens@redhat.com> # # Copyright 2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.errors import * from pykickstart.options import * import gettext _ = lambda x: gettext.ldgettext("pykickstart", x) class FC4_MediaCheck(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.op = self._getParser() self.mediacheck = kwargs.get("mediacheck", False) def __str__(self): retval = KickstartCommand.__str__(self) if self.mediacheck: retval += "mediacheck\n" return retval def _getParser(self): op = KSOptionParser() return op def parse(self, args): (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) if len(extra) > 0: raise KickstartValueError, formatErrorMsg(self.lineno, msg=_("Kickstart command %s does not take any arguments") % "mediacheck") self.mediacheck = True return self ########NEW FILE######## __FILENAME__ = method # # Chris Lumens <clumens@redhat.com> # # Copyright 2007, 2009 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.errors import * from pykickstart.options import * import gettext _ = lambda x: gettext.ldgettext("pykickstart", x) class FC3_Method(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.method = kwargs.get("method", "") # Set all these attributes so calls to this command's __call__ # method can set them. However we don't want to provide them as # arguments to __init__ because method is special. self.biospart = None self.partition = None self.server = None self.dir = None self.url = None def __str__(self): retval = KickstartCommand.__str__(self) if self.method == "cdrom": retval += "# Use CDROM installation media\ncdrom\n" elif self.method == "harddrive": msg = "# Use hard drive installation media\nharddrive --dir=%s" % self.dir if self.biospart is not None: retval += msg + " --biospart=%s\n" % self.biospart else: retval += msg + " --partition=%s\n" % self.partition elif self.method == "nfs": retval += "# Use NFS installation media\nnfs --server=%s --dir=%s\n" % (self.server, self.dir) elif self.method == "url": retval += "# Use network installation\nurl --url=\"%s\"\n" % self.url return retval def _getParser(self): op = KSOptionParser() # method = "cdrom" falls through to the return if self.currentCmd == "harddrive": op.add_option("--biospart", dest="biospart") op.add_option("--partition", dest="partition") op.add_option("--dir", dest="dir", required=1) elif self.currentCmd == "nfs": op.add_option("--server", dest="server", required=1) op.add_option("--dir", dest="dir", required=1) elif self.currentCmd == "url": op.add_option("--url", dest="url", required=1) return op def parse(self, args): self.method = self.currentCmd op = self._getParser() (opts, extra) = op.parse_args(args=args, lineno=self.lineno) self._setToSelf(op, opts) if self.currentCmd == "harddrive": if self.biospart is None and self.partition is None or \ self.biospart is not None and self.partition is not None: raise KickstartValueError, formatErrorMsg(self.lineno, msg=_("One of biospart or partition options must be specified.")) return self class FC6_Method(FC3_Method): removedKeywords = FC3_Method.removedKeywords removedAttrs = FC3_Method.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): FC3_Method.__init__(self, writePriority, *args, **kwargs) # Same reason for this attribute as the comment in FC3_Method. self.opts = None def __str__(self): retval = KickstartCommand.__str__(self) if self.method == "cdrom": retval += "# Use CDROM installation media\ncdrom\n" elif self.method == "harddrive": msg = "# Use hard drive installation media\nharddrive --dir=%s" % self.dir if self.biospart is not None: retval += msg + " --biospart=%s\n" % self.biospart else: retval += msg + " --partition=%s\n" % self.partition elif self.method == "nfs": retval += "# Use NFS installation media\nnfs --server=%s --dir=%s" % (self.server, self.dir) if self.opts is not None: retval += " --opts=\"%s\"" % self.opts retval += "\n" elif self.method == "url": retval += "# Use network installation\nurl --url=\"%s\"\n" % self.url return retval def _getParser(self): op = FC3_Method._getParser(self) if self.currentCmd == "nfs": op.add_option("--opts", dest="opts") return op class F13_Method(FC6_Method): removedKeywords = FC6_Method.removedKeywords removedAttrs = FC6_Method.removedAttrs def __init__(self, *args, **kwargs): FC6_Method.__init__(self, *args, **kwargs) # And same as all the other __init__ methods. self.proxy = "" def __str__(self): retval = FC6_Method.__str__(self) if self.method == "url" and self.proxy: retval = retval.strip() retval += " --proxy=\"%s\"\n" % self.proxy return retval def _getParser(self): op = FC6_Method._getParser(self) if self.currentCmd == "url": op.add_option("--proxy") return op class F14_Method(F13_Method): removedKeywords = F13_Method.removedKeywords removedAttrs = F13_Method.removedAttrs def __init__(self, *args, **kwargs): F13_Method.__init__(self, *args, **kwargs) self.noverifyssl = False def __str__(self): retval = F13_Method.__str__(self) if self.method == "url" and self.noverifyssl: retval = retval.strip() retval += " --noverifyssl\n" return retval def _getParser(self): op = F13_Method._getParser(self) if self.currentCmd == "url": op.add_option("--noverifyssl", action="store_true", default=False) return op RHEL6_Method = F14_Method ########NEW FILE######## __FILENAME__ = monitor # # Chris Lumens <clumens@redhat.com> # # Copyright 2005, 2006, 2007, 2008 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.errors import * from pykickstart.options import * import gettext _ = lambda x: gettext.ldgettext("pykickstart", x) class FC3_Monitor(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.op = self._getParser() self.hsync = kwargs.get("hsync", "") self.monitor = kwargs.get("monitor", "") self.vsync = kwargs.get("vsync", "") def __str__(self): retval = KickstartCommand.__str__(self) retval += "monitor" if self.hsync != "": retval += " --hsync=%s" % self.hsync if self.monitor != "": retval += " --monitor=\"%s\"" % self.monitor if self.vsync != "": retval += " --vsync=%s" % self.vsync if retval != "monitor": return retval + "\n" else: return "" def _getParser(self): op = KSOptionParser() op.add_option("--hsync") op.add_option("--monitor") op.add_option("--vsync") return op def parse(self, args): (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) if extra: mapping = {"cmd": "monitor", "options": extra} raise KickstartValueError, formatErrorMsg(self.lineno, msg=_("Unexpected arguments to %(cmd)s command: %(options)s") % mapping) self._setToSelf(self.op, opts) return self class FC6_Monitor(FC3_Monitor): removedKeywords = FC3_Monitor.removedKeywords removedAttrs = FC3_Monitor.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): FC3_Monitor.__init__(self, writePriority, *args, **kwargs) self.probe = kwargs.get("probe", True) def __str__(self): retval = KickstartCommand.__str__(self) retval += "monitor" if self.hsync != "": retval += " --hsync=%s" % self.hsync if self.monitor != "": retval += " --monitor=\"%s\"" % self.monitor if not self.probe: retval += " --noprobe" if self.vsync != "": retval += " --vsync=%s" % self.vsync if retval != "monitor": return retval + "\n" else: return "" def _getParser(self): op = FC3_Monitor._getParser(self) op.add_option("--noprobe", dest="probe", action="store_false", default=True) return op class F10_Monitor(DeprecatedCommand): def __init__(self): DeprecatedCommand.__init__(self) ########NEW FILE######## __FILENAME__ = mouse # # Chris Lumens <clumens@redhat.com> # # Copyright 2005, 2006, 2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.errors import * from pykickstart.options import * import gettext _ = lambda x: gettext.ldgettext("pykickstart", x) class RHEL3_Mouse(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.op = self._getParser() self.device = kwargs.get("device", "") self.emulthree = kwargs.get("emulthree", False) self.mouse = kwargs.get("mouse", "") def __str__(self): retval = KickstartCommand.__str__(self) opts = "" if self.device: opts += "--device=%s " % self.device if self.emulthree: opts += "--emulthree " if self.mouse: retval += "# System mouse\nmouse %s%s\n" % (opts, self.mouse) return retval def _getParser(self): op = KSOptionParser() op.add_option("--device", dest="device", default="") op.add_option("--emulthree", dest="emulthree", default=False, action="store_true") return op def parse(self, args): (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) self._setToSelf(self.op, opts) if len(extra) != 1: raise KickstartValueError, formatErrorMsg(self.lineno, msg=_("Kickstart command %s requires one argument") % "mouse") self.mouse = extra[0] return self class FC3_Mouse(DeprecatedCommand): def __init__(self): DeprecatedCommand.__init__(self) ########NEW FILE######## __FILENAME__ = multipath # # Chris Lumens <clumens@redhat.com> # Peter Jones <pjones@redhat.com> # # Copyright 2006, 2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.errors import * from pykickstart.options import * import gettext _ = lambda x: gettext.ldgettext("pykickstart", x) class FC6_MpPathData(BaseData): removedKeywords = BaseData.removedKeywords removedAttrs = BaseData.removedAttrs def __init__(self, *args, **kwargs): BaseData.__init__(self, *args, **kwargs) self.mpdev = kwargs.get("mpdev", "") self.device = kwargs.get("device", "") self.rule = kwargs.get("rule", "") def __str__(self): return " --device=%s --rule=\"%s\"" % (self.device, self.rule) class FC6_MultiPathData(BaseData): removedKeywords = BaseData.removedKeywords removedAttrs = BaseData.removedAttrs def __init__(self, *args, **kwargs): BaseData.__init__(self, *args, **kwargs) self.name = kwargs.get("name", "") self.paths = kwargs.get("paths", []) def __str__(self): retval = BaseData.__str__(self) for path in self.paths: retval += "multipath --mpdev=%s %s\n" % (self.name, path.__str__()) return retval class FC6_MultiPath(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=50, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.op = self._getParser() self.mpaths = kwargs.get("mpaths", []) def __str__(self): retval = "" for mpath in self.mpaths: retval += mpath.__str__() return retval def _getParser(self): op = KSOptionParser() op.add_option("--name", dest="name", action="store", type="string", required=1) op.add_option("--device", dest="device", action="store", type="string", required=1) op.add_option("--rule", dest="rule", action="store", type="string", required=1) return op def parse(self, args): (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) dd = FC6_MpPathData() self._setToObj(self.op, opts, dd) dd.lineno = self.lineno dd.mpdev = dd.mpdev.split('/')[-1] parent = None for x in range(0, len(self.mpaths)): mpath = self.mpaths[x] for path in mpath.paths: if path.device == dd.device: mapping = {"device": path.device, "multipathdev": path.mpdev} raise KickstartValueError, formatErrorMsg(self.lineno, msg=_("Device '%(device)s' is already used in multipath '%(multipathdev)s'") % mapping) if mpath.name == dd.mpdev: parent = x if parent is None: mpath = FC6_MultiPathData() return mpath else: mpath = self.mpaths[parent] return dd def dataList(self): return self.mpaths ########NEW FILE######## __FILENAME__ = network # # Chris Lumens <clumens@redhat.com> # # Copyright 2005, 2006, 2007, 2008 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.constants import * from pykickstart.errors import * from pykickstart.options import * import gettext import warnings _ = lambda x: gettext.ldgettext("pykickstart", x) class FC3_NetworkData(BaseData): removedKeywords = BaseData.removedKeywords removedAttrs = BaseData.removedAttrs def __init__(self, *args, **kwargs): BaseData.__init__(self, *args, **kwargs) self.bootProto = kwargs.get("bootProto", BOOTPROTO_DHCP) self.dhcpclass = kwargs.get("dhcpclass", "") self.device = kwargs.get("device", "") self.essid = kwargs.get("essid", "") self.ethtool = kwargs.get("ethtool", "") self.gateway = kwargs.get("gateway", "") self.hostname = kwargs.get("hostname", "") self.ip = kwargs.get("ip", "") self.mtu = kwargs.get("mtu", "") self.nameserver = kwargs.get("nameserver", "") self.netmask = kwargs.get("netmask", "") self.nodns = kwargs.get("nodns", False) self.onboot = kwargs.get("onboot", True) self.wepkey = kwargs.get("wepkey", "") def __eq__(self, y): return self.device and self.device == y.device def _getArgsAsStr(self): retval = "" if self.bootProto != "": retval += " --bootproto=%s" % self.bootProto if self.dhcpclass != "": retval += " --dhcpclass=%s" % self.dhcpclass if self.device != "": retval += " --device=%s" % self.device if self.essid != "": retval += " --essid=\"%s\"" % self.essid if self.ethtool != "": retval += " --ethtool=\"%s\"" % self.ethtool if self.gateway != "": retval += " --gateway=%s" % self.gateway if self.hostname != "": retval += " --hostname=%s" % self.hostname if self.ip != "": retval += " --ip=%s" % self.ip if self.mtu != "": retval += " --mtu=%s" % self.mtu if self.nameserver != "": retval += " --nameserver=%s" % self.nameserver if self.netmask != "": retval += " --netmask=%s" % self.netmask if self.nodns: retval += " --nodns" if not self.onboot: retval += " --onboot=off" if self.wepkey != "": retval += " --wepkey=%s" % self.wepkey return retval def __str__(self): retval = BaseData.__str__(self) retval += "network %s\n" % self._getArgsAsStr() return retval class FC4_NetworkData(FC3_NetworkData): removedKeywords = FC3_NetworkData.removedKeywords removedAttrs = FC3_NetworkData.removedAttrs def __init__(self, *args, **kwargs): FC3_NetworkData.__init__(self, *args, **kwargs) self.notksdevice = kwargs.get("notksdevice", False) def _getArgsAsStr(self): retval = FC3_NetworkData._getArgsAsStr(self) if self.notksdevice: retval += " --notksdevice" return retval class FC6_NetworkData(FC4_NetworkData): removedKeywords = FC4_NetworkData.removedKeywords removedAttrs = FC4_NetworkData.removedAttrs def __init__(self, *args, **kwargs): FC4_NetworkData.__init__(self, *args, **kwargs) self.noipv4 = kwargs.get("noipv4", False) self.noipv6 = kwargs.get("noipv6", False) def _getArgsAsStr(self): retval = FC4_NetworkData._getArgsAsStr(self) if self.noipv4: retval += " --noipv4" if self.noipv6: retval += " --noipv6" return retval class F8_NetworkData(FC6_NetworkData): removedKeywords = FC6_NetworkData.removedKeywords removedAttrs = FC6_NetworkData.removedAttrs def __init__(self, *args, **kwargs): FC6_NetworkData.__init__(self, *args, **kwargs) self.ipv6 = kwargs.get("ipv6", "") def _getArgsAsStr(self): retval = FC6_NetworkData._getArgsAsStr(self) if self.ipv6 != "": retval += " --ipv6" % self.ipv6 return retval class F16_NetworkData(F8_NetworkData): removedKeywords = F8_NetworkData.removedKeywords removedAttrs = F8_NetworkData.removedAttrs def __init__(self, *args, **kwargs): F8_NetworkData.__init__(self, *args, **kwargs) self.activate = kwargs.get("activate", False) self.nodefroute = kwargs.get("nodefroute", False) self.wpakey = kwargs.get("wpakey", "") def _getArgsAsStr(self): retval = F8_NetworkData._getArgsAsStr(self) if self.activate: retval += " --activate" if self.nodefroute: retval += " --nodefroute" if self.wpakey != "": retval += "--wpakey=%s" % self.wpakey return retval class RHEL4_NetworkData(FC3_NetworkData): removedKeywords = FC3_NetworkData.removedKeywords removedAttrs = FC3_NetworkData.removedAttrs def __init__(self, *args, **kwargs): FC3_NetworkData.__init__(self, *args, **kwargs) self.notksdevice = kwargs.get("notksdevice", False) def _getArgsAsStr(self): retval = FC3_NetworkData._getArgsAsStr(self) if self.notksdevice: retval += " --notksdevice" return retval class RHEL6_NetworkData(F8_NetworkData): removedKeywords = F8_NetworkData.removedKeywords removedAttrs = F8_NetworkData.removedAttrs def __init__(self, *args, **kwargs): F8_NetworkData.__init__(self, *args, **kwargs) self.activate = kwargs.get("activate", False) self.nodefroute = kwargs.get("nodefroute", False) def _getArgsAsStr(self): retval = F8_NetworkData._getArgsAsStr(self) if self.activate: retval += " --activate" if self.nodefroute: retval += " --nodefroute" return retval class FC3_Network(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.bootprotoList = [BOOTPROTO_DHCP, BOOTPROTO_BOOTP, BOOTPROTO_STATIC] self.op = self._getParser() self.network = kwargs.get("network", []) def __str__(self): retval = "" for nic in self.network: retval += nic.__str__() if retval != "": return "# Network information\n" + retval else: return "" def _getParser(self): op = KSOptionParser() op.add_option("--bootproto", dest="bootProto", default=BOOTPROTO_DHCP, choices=self.bootprotoList) op.add_option("--dhcpclass", dest="dhcpclass") op.add_option("--device", dest="device") op.add_option("--essid", dest="essid") op.add_option("--ethtool", dest="ethtool") op.add_option("--gateway", dest="gateway") op.add_option("--hostname", dest="hostname") op.add_option("--ip", dest="ip") op.add_option("--mtu", dest="mtu") op.add_option("--nameserver", dest="nameserver") op.add_option("--netmask", dest="netmask") op.add_option("--nodns", dest="nodns", action="store_true", default=False) op.add_option("--onboot", dest="onboot", action="store", type="ksboolean") op.add_option("--wepkey", dest="wepkey") return op def parse(self, args): (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) nd = self.handler.NetworkData() self._setToObj(self.op, opts, nd) nd.lineno = self.lineno # Check for duplicates in the data list. if nd in self.dataList(): warnings.warn(_("A network device with the name %s has already been defined.") % nd.device) return nd def dataList(self): return self.network class FC4_Network(FC3_Network): removedKeywords = FC3_Network.removedKeywords removedAttrs = FC3_Network.removedAttrs def _getParser(self): op = FC3_Network._getParser(self) op.add_option("--notksdevice", dest="notksdevice", action="store_true", default=False) return op class FC6_Network(FC4_Network): removedKeywords = FC4_Network.removedKeywords removedAttrs = FC4_Network.removedAttrs def _getParser(self): op = FC4_Network._getParser(self) op.add_option("--noipv4", dest="noipv4", action="store_true", default=False) op.add_option("--noipv6", dest="noipv6", action="store_true", default=False) return op class F8_Network(FC6_Network): removedKeywords = FC6_Network.removedKeywords removedAttrs = FC6_Network.removedAttrs def _getParser(self): op = FC6_Network._getParser(self) op.add_option("--ipv6", dest="ipv6") return op class F9_Network(F8_Network): removedKeywords = F8_Network.removedKeywords removedAttrs = F8_Network.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): F8_Network.__init__(self, writePriority, *args, **kwargs) self.bootprotoList.append(BOOTPROTO_QUERY) def _getParser(self): op = F8_Network._getParser(self) op.add_option("--bootproto", dest="bootProto", default=BOOTPROTO_DHCP, choices=self.bootprotoList) return op class F16_Network(F9_Network): removedKeywords = F9_Network.removedKeywords removedAttrs = F9_Network.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): F9_Network.__init__(self, writePriority, *args, **kwargs) self.bootprotoList.append(BOOTPROTO_IBFT) def _getParser(self): op = F9_Network._getParser(self) op.add_option("--activate", dest="activate", action="store_true", default=False) op.add_option("--nodefroute", dest="nodefroute", action="store_true", default=False) op.add_option("--wpakey", dest="wpakey", action="store", default="") return op class RHEL4_Network(FC3_Network): removedKeywords = FC3_Network.removedKeywords removedAttrs = FC3_Network.removedAttrs def _getParser(self): op = FC3_Network._getParser(self) op.add_option("--notksdevice", dest="notksdevice", action="store_true", default=False) return op class RHEL5_Network(FC6_Network): removedKeywords = FC6_Network.removedKeywords removedAttrs = FC6_Network.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): FC6_Network.__init__(self, writePriority, *args, **kwargs) self.bootprotoList.append(BOOTPROTO_QUERY) def _getParser(self): op = FC6_Network._getParser(self) op.add_option("--bootproto", dest="bootProto", default=BOOTPROTO_DHCP, choices=self.bootprotoList) return op class RHEL6_Network(F9_Network): removedKeywords = F9_Network.removedKeywords removedAttrs = F9_Network.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): F9_Network.__init__(self, writePriority, *args, **kwargs) self.bootprotoList.append(BOOTPROTO_IBFT) def _getParser(self): op = F9_Network._getParser(self) op.add_option("--activate", dest="activate", action="store_true", default=False) op.add_option("--nodefroute", dest="nodefroute", action="store_true", default=False) return op ########NEW FILE######## __FILENAME__ = partition # # Chris Lumens <clumens@redhat.com> # # Copyright 2005, 2006, 2007, 2008 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.errors import * from pykickstart.options import * import gettext import warnings _ = lambda x: gettext.ldgettext("pykickstart", x) class FC3_PartData(BaseData): removedKeywords = BaseData.removedKeywords removedAttrs = BaseData.removedAttrs def __init__(self, *args, **kwargs): BaseData.__init__(self, *args, **kwargs) self.active = kwargs.get("active", False) self.primOnly = kwargs.get("primOnly", False) self.end = kwargs.get("end", 0) self.fstype = kwargs.get("fstype", "") self.grow = kwargs.get("grow", False) self.maxSizeMB = kwargs.get("maxSizeMB", 0) self.format = kwargs.get("format", True) self.onbiosdisk = kwargs.get("onbiosdisk", "") self.disk = kwargs.get("disk", "") self.onPart = kwargs.get("onPart", "") self.recommended = kwargs.get("recommended", False) self.size = kwargs.get("size", None) self.start = kwargs.get("start", 0) self.mountpoint = kwargs.get("mountpoint", "") def __eq__(self, y): if self.mountpoint: return self.mountpoint == y.mountpoint else: return False def _getArgsAsStr(self): retval = "" if self.active: retval += " --active" if self.primOnly: retval += " --asprimary" if hasattr(self, "end") and self.end != 0: retval += " --end=%s" % self.end if self.fstype != "": retval += " --fstype=\"%s\"" % self.fstype if self.grow: retval += " --grow" if self.maxSizeMB > 0: retval += " --maxsize=%d" % self.maxSizeMB if not self.format: retval += " --noformat" if self.onbiosdisk != "": retval += " --onbiosdisk=%s" % self.onbiosdisk if self.disk != "": retval += " --ondisk=%s" % self.disk if self.onPart != "": retval += " --onpart=%s" % self.onPart if self.recommended: retval += " --recommended" if self.size and self.size != 0: retval += " --size=%s" % self.size if hasattr(self, "start") and self.start != 0: retval += " --start=%s" % self.start return retval def __str__(self): retval = BaseData.__str__(self) if self.mountpoint: mountpoint_str = "%s" % self.mountpoint else: mountpoint_str = "(No mount point)" retval += "part %s%s\n" % (mountpoint_str, self._getArgsAsStr()) return retval class FC4_PartData(FC3_PartData): removedKeywords = FC3_PartData.removedKeywords removedAttrs = FC3_PartData.removedAttrs def __init__(self, *args, **kwargs): FC3_PartData.__init__(self, *args, **kwargs) self.bytesPerInode = kwargs.get("bytesPerInode", 4096) self.fsopts = kwargs.get("fsopts", "") self.label = kwargs.get("label", "") def _getArgsAsStr(self): retval = FC3_PartData._getArgsAsStr(self) if hasattr(self, "bytesPerInode") and self.bytesPerInode != 0: retval += " --bytes-per-inode=%d" % self.bytesPerInode if self.fsopts != "": retval += " --fsoptions=\"%s\"" % self.fsopts if self.label != "": retval += " --label=%s" % self.label return retval class RHEL5_PartData(FC4_PartData): removedKeywords = FC4_PartData.removedKeywords removedAttrs = FC4_PartData.removedAttrs def __init__(self, *args, **kwargs): FC4_PartData.__init__(self, *args, **kwargs) self.encrypted = kwargs.get("encrypted", False) self.passphrase = kwargs.get("passphrase", "") def _getArgsAsStr(self): retval = FC4_PartData._getArgsAsStr(self) if self.encrypted: retval += " --encrypted" if self.passphrase != "": retval += " --passphrase=\"%s\"" % self.passphrase return retval class F9_PartData(FC4_PartData): removedKeywords = FC4_PartData.removedKeywords + ["bytesPerInode"] removedAttrs = FC4_PartData.removedAttrs + ["bytesPerInode"] def __init__(self, *args, **kwargs): FC4_PartData.__init__(self, *args, **kwargs) self.deleteRemovedAttrs() self.fsopts = kwargs.get("fsopts", "") self.label = kwargs.get("label", "") self.fsprofile = kwargs.get("fsprofile", "") self.encrypted = kwargs.get("encrypted", False) self.passphrase = kwargs.get("passphrase", "") def _getArgsAsStr(self): retval = FC4_PartData._getArgsAsStr(self) if self.fsprofile != "": retval += " --fsprofile=\"%s\"" % self.fsprofile if self.encrypted: retval += " --encrypted" if self.passphrase != "": retval += " --passphrase=\"%s\"" % self.passphrase return retval class F11_PartData(F9_PartData): removedKeywords = F9_PartData.removedKeywords + ["start", "end"] removedAttrs = F9_PartData.removedAttrs + ["start", "end"] class F12_PartData(F11_PartData): removedKeywords = F11_PartData.removedKeywords removedAttrs = F11_PartData.removedAttrs def __init__(self, *args, **kwargs): F11_PartData.__init__(self, *args, **kwargs) self.escrowcert = kwargs.get("escrowcert", "") self.backuppassphrase = kwargs.get("backuppassphrase", False) def _getArgsAsStr(self): retval = F11_PartData._getArgsAsStr(self) if self.encrypted and self.escrowcert != "": retval += " --escrowcert=\"%s\"" % self.escrowcert if self.backuppassphrase: retval += " --backuppassphrase" return retval F14_PartData = F12_PartData class FC3_Partition(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=130, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.op = self._getParser() self.partitions = kwargs.get("partitions", []) def __str__(self): retval = "" for part in self.partitions: retval += part.__str__() if retval != "": return "# Disk partitioning information\n" + retval else: return "" def _getParser(self): def part_cb (option, opt_str, value, parser): if value.startswith("/dev/"): parser.values.ensure_value(option.dest, value[5:]) else: parser.values.ensure_value(option.dest, value) op = KSOptionParser() op.add_option("--active", dest="active", action="store_true", default=False) op.add_option("--asprimary", dest="primOnly", action="store_true", default=False) op.add_option("--end", dest="end", action="store", type="int", nargs=1) op.add_option("--fstype", "--type", dest="fstype") op.add_option("--grow", dest="grow", action="store_true", default=False) op.add_option("--maxsize", dest="maxSizeMB", action="store", type="int", nargs=1) op.add_option("--noformat", dest="format", action="store_false", default=True) op.add_option("--onbiosdisk", dest="onbiosdisk") op.add_option("--ondisk", "--ondrive", dest="disk") op.add_option("--onpart", "--usepart", dest="onPart", action="callback", callback=part_cb, nargs=1, type="string") op.add_option("--recommended", dest="recommended", action="store_true", default=False) op.add_option("--size", dest="size", action="store", type="int", nargs=1) op.add_option("--start", dest="start", action="store", type="int", nargs=1) return op def parse(self, args): (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) pd = self.handler.PartData() self._setToObj(self.op, opts, pd) pd.lineno = self.lineno if extra: pd.mountpoint = extra[0] if pd in self.dataList(): warnings.warn(_("A partition with the mountpoint %s has already been defined.") % pd.mountpoint) else: pd.mountpoint = None return pd def dataList(self): return self.partitions class FC4_Partition(FC3_Partition): removedKeywords = FC3_Partition.removedKeywords removedAttrs = FC3_Partition.removedAttrs def __init__(self, writePriority=130, *args, **kwargs): FC3_Partition.__init__(self, writePriority, *args, **kwargs) def part_cb (option, opt_str, value, parser): if value.startswith("/dev/"): parser.values.ensure_value(option.dest, value[5:]) else: parser.values.ensure_value(option.dest, value) def _getParser(self): op = FC3_Partition._getParser(self) op.add_option("--bytes-per-inode", dest="bytesPerInode", action="store", type="int", nargs=1) op.add_option("--fsoptions", dest="fsopts") op.add_option("--label", dest="label") return op class RHEL5_Partition(FC4_Partition): removedKeywords = FC4_Partition.removedKeywords removedAttrs = FC4_Partition.removedAttrs def __init__(self, writePriority=130, *args, **kwargs): FC4_Partition.__init__(self, writePriority, *args, **kwargs) def part_cb (option, opt_str, value, parser): if value.startswith("/dev/"): parser.values.ensure_value(option.dest, value[5:]) else: parser.values.ensure_value(option.dest, value) def _getParser(self): op = FC4_Partition._getParser(self) op.add_option("--encrypted", action="store_true", default=False) op.add_option("--passphrase") return op class F9_Partition(FC4_Partition): removedKeywords = FC4_Partition.removedKeywords removedAttrs = FC4_Partition.removedAttrs def __init__(self, writePriority=130, *args, **kwargs): FC4_Partition.__init__(self, writePriority, *args, **kwargs) def part_cb (option, opt_str, value, parser): if value.startswith("/dev/"): parser.values.ensure_value(option.dest, value[5:]) else: parser.values.ensure_value(option.dest, value) def _getParser(self): op = FC4_Partition._getParser(self) op.add_option("--bytes-per-inode", deprecated=1) op.add_option("--fsprofile") op.add_option("--encrypted", action="store_true", default=False) op.add_option("--passphrase") return op class F11_Partition(F9_Partition): removedKeywords = F9_Partition.removedKeywords removedAttrs = F9_Partition.removedAttrs def _getParser(self): op = F9_Partition._getParser(self) op.add_option("--start", deprecated=1) op.add_option("--end", deprecated=1) return op class F12_Partition(F11_Partition): removedKeywords = F11_Partition.removedKeywords removedAttrs = F11_Partition.removedAttrs def _getParser(self): op = F11_Partition._getParser(self) op.add_option("--escrowcert") op.add_option("--backuppassphrase", action="store_true", default=False) return op class F14_Partition(F12_Partition): removedKeywords = F12_Partition.removedKeywords removedAttrs = F12_Partition.removedAttrs def _getParser(self): op = F12_Partition._getParser(self) op.remove_option("--bytes-per-inode") op.remove_option("--start") op.remove_option("--end") return op ########NEW FILE######## __FILENAME__ = raid # # Chris Lumens <clumens@redhat.com> # # Copyright 2005, 2006, 2007, 2008, 2011 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.errors import * from pykickstart.options import * import gettext import warnings _ = lambda x: gettext.ldgettext("pykickstart", x) class FC3_RaidData(BaseData): removedKeywords = BaseData.removedKeywords removedAttrs = BaseData.removedAttrs def __init__(self, *args, **kwargs): BaseData.__init__(self, *args, **kwargs) self.device = kwargs.get("device", None) self.fstype = kwargs.get("fstype", "") self.level = kwargs.get("level", "") self.format = kwargs.get("format", True) self.spares = kwargs.get("spares", 0) self.preexist = kwargs.get("preexist", False) self.mountpoint = kwargs.get("mountpoint", "") self.members = kwargs.get("members", []) def __eq__(self, y): return self.device == y.device def _getArgsAsStr(self): retval = "" if self.device != "": retval += " --device=%s" % self.device if self.fstype != "": retval += " --fstype=\"%s\"" % self.fstype if self.level != "": retval += " --level=%s" % self.level if not self.format: retval += " --noformat" if self.spares != 0: retval += " --spares=%d" % self.spares if self.preexist: retval += " --useexisting" return retval def __str__(self): retval = BaseData.__str__(self) retval += "raid %s%s %s\n" % (self.mountpoint, self._getArgsAsStr(), " ".join(self.members)) return retval class FC4_RaidData(FC3_RaidData): removedKeywords = FC3_RaidData.removedKeywords removedAttrs = FC3_RaidData.removedAttrs def __init__(self, *args, **kwargs): FC3_RaidData.__init__(self, *args, **kwargs) self.fsopts = kwargs.get("fsopts", "") def _getArgsAsStr(self): retval = FC3_RaidData._getArgsAsStr(self) if self.fsopts != "": retval += " --fsoptions=\"%s\"" % self.fsopts return retval class FC5_RaidData(FC4_RaidData): removedKeywords = FC4_RaidData.removedKeywords removedAttrs = FC4_RaidData.removedAttrs def __init__(self, *args, **kwargs): FC4_RaidData.__init__(self, *args, **kwargs) self.bytesPerInode = kwargs.get("bytesPerInode", 4096) def _getArgsAsStr(self): retval = FC4_RaidData._getArgsAsStr(self) if hasattr(self, "bytesPerInode") and self.bytesPerInode != 0: retval += " --bytes-per-inode=%d" % self.bytesPerInode return retval class RHEL5_RaidData(FC5_RaidData): removedKeywords = FC5_RaidData.removedKeywords removedAttrs = FC5_RaidData.removedAttrs def __init__(self, *args, **kwargs): FC5_RaidData.__init__(self, *args, **kwargs) self.encrypted = kwargs.get("encrypted", False) self.passphrase = kwargs.get("passphrase", "") def _getArgsAsStr(self): retval = FC5_RaidData._getArgsAsStr(self) if self.encrypted: retval += " --encrypted" if self.passphrase != "": retval += " --passphrase=\"%s\"" % self.passphrase return retval F7_RaidData = FC5_RaidData class F9_RaidData(FC5_RaidData): removedKeywords = FC5_RaidData.removedKeywords + ["bytesPerInode"] removedAttrs = FC5_RaidData.removedAttrs + ["bytesPerInode"] def __init__(self, *args, **kwargs): FC5_RaidData.__init__(self, *args, **kwargs) self.deleteRemovedAttrs() self.fsprofile = kwargs.get("fsprofile", "") self.encrypted = kwargs.get("encrypted", False) self.passphrase = kwargs.get("passphrase", "") def _getArgsAsStr(self): retval = FC5_RaidData._getArgsAsStr(self) if self.fsprofile != "": retval += " --fsprofile=\"%s\"" % self.fsprofile if self.encrypted: retval += " --encrypted" if self.passphrase != "": retval += " --passphrase=\"%s\"" % self.passphrase return retval class F12_RaidData(F9_RaidData): removedKeywords = F9_RaidData.removedKeywords removedAttrs = F9_RaidData.removedAttrs def __init__(self, *args, **kwargs): F9_RaidData.__init__(self, *args, **kwargs) self.deleteRemovedAttrs() self.escrowcert = kwargs.get("escrowcert", "") self.backuppassphrase = kwargs.get("backuppassphrase", False) def _getArgsAsStr(self): retval = F9_RaidData._getArgsAsStr(self) if self.encrypted and self.escrowcert != "": retval += " --escrowcert=\"%s\"" % self.escrowcert if self.backuppassphrase: retval += " --backuppassphrase" return retval F13_RaidData = F12_RaidData F14_RaidData = F13_RaidData class F15_RaidData(F14_RaidData): removedKeywords = F14_RaidData.removedKeywords removedAttrs = F14_RaidData.removedAttrs def __init__(self, *args, **kwargs): F14_RaidData.__init__(self, *args, **kwargs) self.deleteRemovedAttrs() self.label = kwargs.get("label", "") def _getArgsAsStr(self): retval = F14_RaidData._getArgsAsStr(self) if self.label != "": retval += " --label=%s" % self.label return retval class FC3_Raid(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=131, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.op = self._getParser() # A dict of all the RAID levels we support. This means that if we # support more levels in the future, subclasses don't have to # duplicate too much. self.levelMap = { "RAID0": "RAID0", "0": "RAID0", "RAID1": "RAID1", "1": "RAID1", "RAID5": "RAID5", "5": "RAID5", "RAID6": "RAID6", "6": "RAID6" } self.raidList = kwargs.get("raidList", []) def __str__(self): retval = "" for raid in self.raidList: retval += raid.__str__() return retval def _getParser(self): def raid_cb (option, opt_str, value, parser): parser.values.format = False parser.values.preexist = True def device_cb (option, opt_str, value, parser): if value[0:2] == "md": parser.values.ensure_value(option.dest, value[2:]) else: parser.values.ensure_value(option.dest, value) def level_cb (option, opt_str, value, parser): if self.levelMap.has_key(value): parser.values.ensure_value(option.dest, self.levelMap[value]) op = KSOptionParser() op.add_option("--device", action="callback", callback=device_cb, dest="device", type="string", nargs=1, required=1) op.add_option("--fstype", dest="fstype") op.add_option("--level", dest="level", action="callback", callback=level_cb, type="string", nargs=1) op.add_option("--noformat", action="callback", callback=raid_cb, dest="format", default=True, nargs=0) op.add_option("--spares", dest="spares", action="store", type="int", nargs=1, default=0) op.add_option("--useexisting", dest="preexist", action="store_true", default=False) return op def parse(self, args): (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) if len(extra) == 0: raise KickstartValueError, formatErrorMsg(self.lineno, msg=_("Mount point required for %s") % "raid") if len(extra) == 1: raise KickstartValueError, formatErrorMsg(self.lineno, msg=_("Partitions required for %s") % "raid") rd = self.handler.RaidData() self._setToObj(self.op, opts, rd) rd.lineno = self.lineno # --device can't just take an int in the callback above, because it # could be specificed as "mdX", which causes optparse to error when # it runs int(). rd.device = int(rd.device) rd.mountpoint = extra[0] rd.members = extra[1:] # Check for duplicates in the data list. if rd in self.dataList(): warnings.warn(_("A RAID device with the name %s has already been defined.") % rd.device) return rd def dataList(self): return self.raidList class FC4_Raid(FC3_Raid): removedKeywords = FC3_Raid.removedKeywords removedAttrs = FC3_Raid.removedAttrs def _getParser(self): op = FC3_Raid._getParser(self) op.add_option("--fsoptions", dest="fsopts") return op class FC5_Raid(FC4_Raid): removedKeywords = FC4_Raid.removedKeywords removedAttrs = FC4_Raid.removedAttrs def _getParser(self): op = FC4_Raid._getParser(self) op.add_option("--bytes-per-inode", dest="bytesPerInode", action="store", type="int", nargs=1) return op class RHEL5_Raid(FC5_Raid): removedKeywords = FC5_Raid.removedKeywords removedAttrs = FC5_Raid.removedAttrs def __init__(self, writePriority=131, *args, **kwargs): FC5_Raid.__init__(self, writePriority, *args, **kwargs) self.levelMap.update({"RAID10": "RAID10", "10": "RAID10"}) def _getParser(self): op = FC5_Raid._getParser(self) op.add_option("--encrypted", action="store_true", default=False) op.add_option("--passphrase") return op class F7_Raid(FC5_Raid): removedKeywords = FC5_Raid.removedKeywords removedAttrs = FC5_Raid.removedAttrs def __init__(self, writePriority=131, *args, **kwargs): FC5_Raid.__init__(self, writePriority, *args, **kwargs) self.levelMap.update({"RAID10": "RAID10", "10": "RAID10"}) class F9_Raid(F7_Raid): removedKeywords = F7_Raid.removedKeywords removedAttrs = F7_Raid.removedAttrs def _getParser(self): op = F7_Raid._getParser(self) op.add_option("--bytes-per-inode", deprecated=1) op.add_option("--fsprofile") op.add_option("--encrypted", action="store_true", default=False) op.add_option("--passphrase") return op class F12_Raid(F9_Raid): removedKeywords = F9_Raid.removedKeywords removedAttrs = F9_Raid.removedAttrs def _getParser(self): op = F9_Raid._getParser(self) op.add_option("--escrowcert") op.add_option("--backuppassphrase", action="store_true", default=False) return op class F13_Raid(F12_Raid): removedKeywords = F12_Raid.removedKeywords removedAttrs = F12_Raid.removedAttrs def __init__(self, writePriority=131, *args, **kwargs): F12_Raid.__init__(self, writePriority, *args, **kwargs) self.levelMap.update({"RAID4": "RAID4", "4": "RAID4"}) class F14_Raid(F13_Raid): removedKeywords = F13_Raid.removedKeywords removedAttrs = F13_Raid.removedAttrs def _getParser(self): op = F13_Raid._getParser(self) op.remove_option("--bytes-per-inode") return op class F15_Raid(F14_Raid): removedKeywords = F14_Raid.removedKeywords removedAttrs = F14_Raid.removedAttrs def _getParser(self): op = F14_Raid._getParser(self) op.add_option("--label") return op ########NEW FILE######## __FILENAME__ = reboot # # Chris Lumens <clumens@redhat.com> # # Copyright 2005, 2006, 2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.constants import * from pykickstart.errors import * from pykickstart.options import * class FC3_Reboot(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.action = kwargs.get("action", None) def __str__(self): retval = KickstartCommand.__str__(self) if self.action == KS_REBOOT: retval += "# Reboot after installation\nreboot\n" elif self.action == KS_SHUTDOWN: retval += "# Shutdown after installation\nshutdown\n" return retval def parse(self, args): if self.currentCmd == "reboot": self.action = KS_REBOOT else: self.action = KS_SHUTDOWN return self class FC6_Reboot(FC3_Reboot): removedKeywords = FC3_Reboot.removedKeywords removedAttrs = FC3_Reboot.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): FC3_Reboot.__init__(self, writePriority, *args, **kwargs) self.op = self._getParser() self.eject = kwargs.get("eject", False) def __str__(self): retval = FC3_Reboot.__str__(self).rstrip() if self.eject: retval += " --eject" return retval + "\n" def _getParser(self): op = KSOptionParser() op.add_option("--eject", dest="eject", action="store_true", default=False) return op def parse(self, args): FC3_Reboot.parse(self, args) (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) self._setToSelf(self.op, opts) return self ########NEW FILE######## __FILENAME__ = repo # # Chris Lumens <clumens@redhat.com> # # Copyright 2007, 2008, 2009 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.constants import * from pykickstart.errors import * from pykickstart.options import * import gettext import warnings _ = lambda x: gettext.ldgettext("pykickstart", x) class FC6_RepoData(BaseData): removedKeywords = BaseData.removedKeywords removedAttrs = BaseData.removedAttrs def __init__(self, *args, **kwargs): BaseData.__init__(self, *args, **kwargs) self.baseurl = kwargs.get("baseurl", "") self.mirrorlist = kwargs.get("mirrorlist", None) self.name = kwargs.get("name", "") def __eq__(self, y): return self.name == y.name def _getArgsAsStr(self): retval = "" if self.baseurl: retval += "--baseurl=%s" % self.baseurl elif self.mirrorlist: retval += "--mirrorlist=%s" % self.mirrorlist return retval def __str__(self): retval = BaseData.__str__(self) retval += "repo --name=\"%s\" %s\n" % (self.name, self._getArgsAsStr()) return retval class F8_RepoData(FC6_RepoData): removedKeywords = FC6_RepoData.removedKeywords removedAttrs = FC6_RepoData.removedAttrs def __init__(self, *args, **kwargs): FC6_RepoData.__init__(self, *args, **kwargs) self.cost = kwargs.get("cost", None) self.includepkgs = kwargs.get("includepkgs", []) self.excludepkgs = kwargs.get("excludepkgs", []) def _getArgsAsStr(self): retval = FC6_RepoData._getArgsAsStr(self) if self.cost: retval += " --cost=%s" % self.cost if self.includepkgs: retval += " --includepkgs=\"%s\"" % ",".join(self.includepkgs) if self.excludepkgs: retval += " --excludepkgs=\"%s\"" % ",".join(self.excludepkgs) return retval class F11_RepoData(F8_RepoData): removedKeywords = F8_RepoData.removedKeywords removedAttrs = F8_RepoData.removedAttrs def __init__(self, *args, **kwargs): F8_RepoData.__init__(self, *args, **kwargs) self.ignoregroups = kwargs.get("ignoregroups", None) def _getArgsAsStr(self): retval = F8_RepoData._getArgsAsStr(self) if self.ignoregroups: retval += " --ignoregroups=true" return retval class F13_RepoData(F11_RepoData): removedKeywords = F11_RepoData.removedKeywords removedAttrs = F11_RepoData.removedAttrs def __init__(self, *args, **kwargs): F11_RepoData.__init__(self, *args, **kwargs) self.proxy = kwargs.get("proxy", "") def _getArgsAsStr(self): retval = F11_RepoData._getArgsAsStr(self) if self.proxy: retval += " --proxy=\"%s\"" % self.proxy return retval class F14_RepoData(F13_RepoData): removedKeywords = F13_RepoData.removedKeywords removedAttrs = F13_RepoData.removedAttrs def __init__(self, *args, **kwargs): F13_RepoData.__init__(self, *args, **kwargs) self.noverifyssl = kwargs.get("noverifyssl", False) def _getArgsAsStr(self): retval = F13_RepoData._getArgsAsStr(self) if self.noverifyssl: retval += " --noverifyssl" return retval RHEL6_RepoData = F14_RepoData F15_RepoData = F14_RepoData class FC6_Repo(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs urlRequired = True def __init__(self, writePriority=0, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.op = self._getParser() self.repoList = kwargs.get("repoList", []) def __str__(self): retval = "" for repo in self.repoList: retval += repo.__str__() return retval def _getParser(self): op = KSOptionParser() op.add_option("--name", dest="name", required=1) op.add_option("--baseurl") op.add_option("--mirrorlist") return op def parse(self, args): (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) if len(extra) != 0: mapping = {"command": "repo", "options": extra} raise KickstartValueError, formatErrorMsg(self.lineno, msg=_("Unexpected arguments to %(command)s command: %(options)s") % mapping) # This is lame, but I can't think of a better way to make sure only # one of these two is specified. if opts.baseurl and opts.mirrorlist: raise KickstartValueError, formatErrorMsg(self.lineno, msg=_("Only one of --baseurl and --mirrorlist may be specified for repo command.")) if self.urlRequired and not opts.baseurl and not opts.mirrorlist: raise KickstartValueError, formatErrorMsg(self.lineno, msg=_("One of --baseurl or --mirrorlist must be specified for repo command.")) rd = self.handler.RepoData() self._setToObj(self.op, opts, rd) rd.lineno = self.lineno # Check for duplicates in the data list. if rd in self.dataList(): warnings.warn(_("A repo with the name %s has already been defined.") % rd.name) return rd def dataList(self): return self.repoList class F8_Repo(FC6_Repo): removedKeywords = FC6_Repo.removedKeywords removedAttrs = FC6_Repo.removedAttrs def __str__(self): retval = "" for repo in self.repoList: retval += repo.__str__() return retval def _getParser(self): def list_cb (option, opt_str, value, parser): for d in value.split(','): parser.values.ensure_value(option.dest, []).append(d) op = FC6_Repo._getParser(self) op.add_option("--cost", action="store", type="int") op.add_option("--excludepkgs", action="callback", callback=list_cb, nargs=1, type="string") op.add_option("--includepkgs", action="callback", callback=list_cb, nargs=1, type="string") return op def methodToRepo(self): if not self.handler.method.url: raise KickstartError, formatErrorMsg(self.handler.method.lineno, msg=_("Method must be a url to be added to the repo list.")) reponame = "ks-method-url" repourl = self.handler.method.url rd = self.handler.RepoData(name=reponame, baseurl=repourl) return rd class F11_Repo(F8_Repo): removedKeywords = F8_Repo.removedKeywords removedAttrs = F8_Repo.removedAttrs def _getParser(self): op = F8_Repo._getParser(self) op.add_option("--ignoregroups", action="store", type="ksboolean") return op class F13_Repo(F11_Repo): removedKeywords = F11_Repo.removedKeywords removedAttrs = F11_Repo.removedAttrs def _getParser(self): op = F11_Repo._getParser(self) op.add_option("--proxy") return op class F14_Repo(F13_Repo): removedKeywords = F13_Repo.removedKeywords removedAttrs = F13_Repo.removedAttrs def _getParser(self): op = F13_Repo._getParser(self) op.add_option("--noverifyssl", action="store_true", default=False) return op RHEL6_Repo = F14_Repo class F15_Repo(F14_Repo): removedKeywords = F14_Repo.removedKeywords removedAttrs = F14_Repo.removedAttrs urlRequired = False ########NEW FILE######## __FILENAME__ = rescue # # Alexander Todorov <atodorov@redhat.com> # # Copyright 2008 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.errors import * from pykickstart.options import * import gettext _ = lambda x: gettext.ldgettext("pykickstart", x) class F10_Rescue(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.op = self._getParser() self.rescue = False self.nomount = kwargs.get("nomount", False) self.romount = kwargs.get("romount", False) def __str__(self): retval = KickstartCommand.__str__(self) if self.rescue: retval += "rescue" if self.nomount: retval += " --nomount" if self.romount: retval += " --romount" retval = "# Start rescue mode\n%s\n" % retval return retval def _getParser(self): op = KSOptionParser() op.add_option("--nomount", dest="nomount", action="store_true", default=False) op.add_option("--romount", dest="romount", action="store_true", default=False) return op def parse(self, args): (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) if opts.nomount and opts.romount: raise KickstartValueError, formatErrorMsg(self.lineno, msg=_("Only one of --nomount and --romount may be specified for rescue command.")) self._setToSelf(self.op, opts) self.rescue = True return self ########NEW FILE######## __FILENAME__ = rootpw # # Chris Lumens <clumens@redhat.com> # # Copyright 2005, 2006, 2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.errors import * from pykickstart.options import * import gettext _ = lambda x: gettext.ldgettext("pykickstart", x) class FC3_RootPw(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.op = self._getParser() self.isCrypted = kwargs.get("isCrypted", False) self.password = kwargs.get("password", "") def _getArgsAsStr(self): retval = "" if self.isCrypted: retval += " --iscrypted" return retval def __str__(self): retval = KickstartCommand.__str__(self) if self.password != "": retval += "# Root password\nrootpw%s %s\n" % (self._getArgsAsStr(), self.password) return retval def _getParser(self): op = KSOptionParser() op.add_option("--iscrypted", dest="isCrypted", action="store_true", default=False) return op def parse(self, args): (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) self._setToSelf(self.op, opts) if len(extra) != 1: raise KickstartValueError, formatErrorMsg(self.lineno, msg=_("A single argument is expected for the %s command") % "rootpw") self.password = extra[0] return self class F8_RootPw(FC3_RootPw): removedKeywords = FC3_RootPw.removedKeywords removedAttrs = FC3_RootPw.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): FC3_RootPw.__init__(self, writePriority, *args, **kwargs) self.lock = kwargs.get("lock", False) def _getArgsAsStr(self): retval = FC3_RootPw._getArgsAsStr(self) if self.lock: retval += " --lock" if not self.isCrypted: retval += " --plaintext" return retval def _getParser(self): op = FC3_RootPw._getParser(self) op.add_option("--lock", dest="lock", action="store_true", default=False) op.add_option("--plaintext", dest="isCrypted", action="store_false") return op ########NEW FILE######## __FILENAME__ = selinux # # Chris Lumens <clumens@redhat.com> # # Copyright 2005, 2006, 2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.constants import * from pykickstart.options import * class FC3_SELinux(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.op = self._getParser() self.selinux = kwargs.get("selinux", None) def __str__(self): retval = KickstartCommand.__str__(self) if not retval and self.selinux is None: return "" retval += "# SELinux configuration\n" if self.selinux == SELINUX_DISABLED: retval += "selinux --disabled\n" elif self.selinux == SELINUX_ENFORCING: retval += "selinux --enforcing\n" elif self.selinux == SELINUX_PERMISSIVE: retval += "selinux --permissive\n" return retval def _getParser(self): op = KSOptionParser() op.add_option("--disabled", dest="selinux", action="store_const", const=SELINUX_DISABLED) op.add_option("--enforcing", dest="selinux", action="store_const", const=SELINUX_ENFORCING) op.add_option("--permissive", dest="selinux", action="store_const", const=SELINUX_PERMISSIVE) return op def parse(self, args): (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) self._setToSelf(self.op, opts) return self ########NEW FILE######## __FILENAME__ = services # # Chris Lumens <clumens@redhat.com> # # Copyright 2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.errors import * from pykickstart.options import * import gettext _ = lambda x: gettext.ldgettext("pykickstart", x) class FC6_Services(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.op = self._getParser() self.disabled = kwargs.get("disabled", []) self.enabled = kwargs.get("enabled", []) def __str__(self): retval = KickstartCommand.__str__(self) args = "" if len(self.disabled) > 0: args += " --disabled=\"%s\"" % ",".join(self.disabled) if len(self.enabled) > 0: args += " --enabled=\"%s\"" % ",".join(self.enabled) if args != "": retval += "# System services\nservices%s\n" % args return retval def _getParser(self): def services_cb (option, opt_str, value, parser): for d in value.split(','): parser.values.ensure_value(option.dest, []).append(d.strip()) op = KSOptionParser() op.add_option("--disabled", dest="disabled", action="callback", callback=services_cb, nargs=1, type="string") op.add_option("--enabled", dest="enabled", action="callback", callback=services_cb, nargs=1, type="string") return op def parse(self, args): (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) self._setToSelf(self.op, opts) if len(self.disabled) == 0 and len(self.enabled) == 0: raise KickstartParseError, formatErrorMsg(self.lineno, msg=_("One of --disabled or --enabled must be provided.")) return self ########NEW FILE######## __FILENAME__ = skipx # # Chris Lumens <clumens@redhat.com> # # Copyright 2005, 2006, 2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.errors import * from pykickstart.options import * import gettext _ = lambda x: gettext.ldgettext("pykickstart", x) class FC3_SkipX(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.op = self._getParser() self.skipx = kwargs.get("skipx", False) def __str__(self): retval = KickstartCommand.__str__(self) if self.skipx: retval += "# Do not configure the X Window System\nskipx\n" return retval def _getParser(self): op = KSOptionParser() return op def parse(self, args): (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) if len(extra) > 0: raise KickstartValueError, formatErrorMsg(self.lineno, msg=_("Kickstart command %s does not take any arguments") % "skipx") self.skipx = True return self ########NEW FILE######## __FILENAME__ = sshpw # # Peter Jones <pjones@redhat.com> # # Copyright 2009 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.errors import * from pykickstart.options import * import gettext _ = lambda x: gettext.ldgettext("pykickstart", x) class F13_SshPwData(BaseData): removedKeywords = BaseData.removedKeywords removedAttrs = BaseData.removedAttrs def __init__(self, *args, **kwargs): BaseData.__init__(self, *args, **kwargs) self.username = kwargs.get("username", None) self.isCrypted = kwargs.get("isCrypted", False) self.password = kwargs.get("password", "") self.lock = kwargs.get("lock", False) def __eq__(self, y): return self.username == y.username def __str__(self): retval = BaseData.__str__(self) retval += "sshpw" retval += self._getArgsAsStr() + '\n' return retval def _getArgsAsStr(self): retval = "" retval += " --username=%s" % self.username if self.lock: retval += " --lock" if self.isCrypted: retval += " --iscrypted" else: retval += " --plaintext" retval += " %s" % self.password return retval class F13_SshPw(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.op = self._getParser() self.sshUserList = kwargs.get("sshUserList", []) def __str__(self): retval = "" for user in self.sshUserList: retval += user.__str__() return retval def _getParser(self): op = KSOptionParser() op.add_option("--username", dest="username", required=True) op.add_option("--iscrypted", dest="isCrypted", action="store_true", default=False) op.add_option("--plaintext", dest="isCrypted", action="store_false") op.add_option("--lock", dest="lock", action="store_true", default=False) return op def parse(self, args): ud = self.handler.SshPwData() (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) self._setToObj(self.op, opts, ud) ud.lineno = self.lineno if len(extra) != 1: raise KickstartValueError, formatErrorMsg(self.lineno, msg=_("A single argument is expected for the %s command") % "sshpw") ud.password = extra[0] if ud in self.dataList(): warnings.warn(_("An ssh user with the name %s has already been defined.") % ud.name) return ud def dataList(self): return self.sshUserList ########NEW FILE######## __FILENAME__ = timezone # # Chris Lumens <clumens@redhat.com> # # Copyright 2005, 2006, 2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.errors import * from pykickstart.options import * import gettext _ = lambda x: gettext.ldgettext("pykickstart", x) class FC3_Timezone(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.op = self._getParser() self.isUtc = kwargs.get("isUtc", False) self.timezone = kwargs.get("timezone", "") def __str__(self): retval = KickstartCommand.__str__(self) if self.timezone != "": if self.isUtc: utc = "--utc" else: utc = "" retval += "# System timezone\ntimezone %s %s\n" %(utc, self.timezone) return retval def _getParser(self): op = KSOptionParser() op.add_option("--utc", dest="isUtc", action="store_true", default=False) return op def parse(self, args): (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) self._setToSelf(self.op, opts) if len(extra) != 1: raise KickstartValueError, formatErrorMsg(self.lineno, msg=_("A single argument is expected for the %s command") % "timezone") self.timezone = extra[0] return self class FC6_Timezone(FC3_Timezone): removedKeywords = FC3_Timezone.removedKeywords removedAttrs = FC3_Timezone.removedAttrs def __str__(self): retval = KickstartCommand.__str__(self) if self.timezone != "": if self.isUtc: utc = "--isUtc" else: utc = "" retval += "# System timezone\ntimezone %s %s\n" %(utc, self.timezone) return retval def _getParser(self): op = FC3_Timezone._getParser(self) op.add_option("--utc", "--isUtc", dest="isUtc", action="store_true", default=False) return op ########NEW FILE######## __FILENAME__ = updates # # Chris Lumens <clumens@redhat.com> # # Copyright 2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.errors import * from pykickstart.options import * import gettext _ = lambda x: gettext.ldgettext("pykickstart", x) class F7_Updates(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.op = self._getParser() self.url = kwargs.get("url", "") def __str__(self): retval = KickstartCommand.__str__(self) if self.url == "floppy": retval += "updates\n" elif self.url != "": retval += "updates %s\n" % self.url return retval def _getParser(self): op = KSOptionParser() return op def parse(self, args): (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) if len(extra) > 1: raise KickstartValueError, formatErrorMsg(self.lineno, msg=_("Kickstart command %s only takes one argument") % "updates") elif len(extra) == 0: self.url = "floppy" else: self.url = extra[0] return self ########NEW FILE######## __FILENAME__ = upgrade # # Chris Lumens <clumens@redhat.com> # # Copyright 2005, 2006, 2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.errors import * from pykickstart.options import * import gettext _ = lambda x: gettext.ldgettext("pykickstart", x) class FC3_Upgrade(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.upgrade = kwargs.get("upgrade", None) self.op = self._getParser() def __str__(self): retval = KickstartCommand.__str__(self) if self.upgrade is None: return retval if self.upgrade: retval += "# Upgrade existing installation\nupgrade\n" else: retval += "# Install OS instead of upgrade\ninstall\n" return retval def _getParser(self): op = KSOptionParser() return op def parse(self, args): (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) if len(extra) > 0: raise KickstartValueError, formatErrorMsg(self.lineno, msg=_("Kickstart command %s does not take any arguments") % "upgrade") if self.currentCmd == "upgrade": self.upgrade = True else: self.upgrade = False return self class F11_Upgrade(FC3_Upgrade): removedKeywords = FC3_Upgrade.removedKeywords removedAttrs = FC3_Upgrade.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): FC3_Upgrade.__init__(self, writePriority, *args, **kwargs) self.op = self._getParser() self.root_device = kwargs.get("root_device", None) def __str__(self): if self.upgrade and (self.root_device is not None): retval = KickstartCommand.__str__(self) retval += "# Upgrade existing installation\nupgrade --root-device=%s\n" % self.root_device else: retval = FC3_Upgrade.__str__(self) return retval def _getParser(self): op = KSOptionParser() op.add_option("--root-device", dest="root_device") return op def parse(self, args): (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) if len(extra) > 0: raise KickstartValueError, formatErrorMsg(self.lineno, msg=_("Kickstart command %s does not take any arguments") % "upgrade") if (opts.root_device is not None) and (opts.root_device == ""): raise KickstartValueError, formatErrorMsg(self.lineno, msg=_("Kickstart command %s does not accept empty parameter %s") % ("upgrade", "--root-device")) else: self.root_device = opts.root_device if self.currentCmd == "upgrade": self.upgrade = True else: self.upgrade = False return self ########NEW FILE######## __FILENAME__ = user # # Chris Lumens <clumens@redhat.com> # # Copyright 2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.constants import * from pykickstart.errors import * from pykickstart.options import * import gettext import warnings _ = lambda x: gettext.ldgettext("pykickstart", x) class FC6_UserData(BaseData): removedKeywords = BaseData.removedKeywords removedAttrs = BaseData.removedAttrs def __init__(self, *args, **kwargs): BaseData.__init__(self, *args, **kwargs) self.groups = kwargs.get("groups", []) self.homedir = kwargs.get("homedir", "") self.isCrypted = kwargs.get("isCrypted", False) self.name = kwargs.get("name", "") self.password = kwargs.get("password", "") self.shell = kwargs.get("shell", "") self.uid = kwargs.get("uid", None) def __eq__(self, y): return self.name == y.name def __str__(self): retval = BaseData.__str__(self) if self.uid != "": retval += "user" retval += self._getArgsAsStr() + "\n" return retval def _getArgsAsStr(self): retval = "" if len(self.groups) > 0: retval += " --groups=%s" % ",".join(self.groups) if self.homedir: retval += " --homedir=%s" % self.homedir if self.name: retval += " --name=%s" % self.name if self.password: retval += " --password=%s" % self.password if self.isCrypted: retval += " --iscrypted" if self.shell: retval += " --shell=%s" % self.shell if self.uid: retval += " --uid=%s" % self.uid return retval class F8_UserData(FC6_UserData): removedKeywords = FC6_UserData.removedKeywords removedAttrs = FC6_UserData.removedAttrs def __init__(self, *args, **kwargs): FC6_UserData.__init__(self, *args, **kwargs) self.lock = kwargs.get("lock", False) def _getArgsAsStr(self): retval = FC6_UserData._getArgsAsStr(self) if self.lock: retval += " --lock" return retval class F12_UserData(F8_UserData): removedKeywords = F8_UserData.removedKeywords removedAttrs = F8_UserData.removedAttrs def __init__(self, *args, **kwargs): F8_UserData.__init__(self, *args, **kwargs) self.gecos = kwargs.get("gecos", "") def _getArgsAsStr(self): retval = F8_UserData._getArgsAsStr(self) if self.gecos: retval += " --gecos=\"%s\"" % (self.gecos,) return retval class FC6_User(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.op = self._getParser() self.userList = kwargs.get("userList", []) def __str__(self): retval = "" for user in self.userList: retval += user.__str__() return retval def _getParser(self): def groups_cb (option, opt_str, value, parser): for d in value.split(','): parser.values.ensure_value(option.dest, []).append(d) op = KSOptionParser() op.add_option("--groups", dest="groups", action="callback", callback=groups_cb, nargs=1, type="string") op.add_option("--homedir") op.add_option("--iscrypted", dest="isCrypted", action="store_true", default=False) op.add_option("--name", required=1) op.add_option("--password") op.add_option("--shell") op.add_option("--uid", type="int") return op def parse(self, args): ud = self.handler.UserData() (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) self._setToObj(self.op, opts, ud) ud.lineno = self.lineno # Check for duplicates in the data list. if ud in self.dataList(): warnings.warn(_("A user with the name %s has already been defined.") % ud.name) return ud def dataList(self): return self.userList class F8_User(FC6_User): removedKeywords = FC6_User.removedKeywords removedAttrs = FC6_User.removedAttrs def _getParser(self): op = FC6_User._getParser(self) op.add_option("--lock", action="store_true", default=False) op.add_option("--plaintext", dest="isCrypted", action="store_false") return op class F12_User(F8_User): removedKeywords = F8_User.removedKeywords removedAttrs = F8_User.removedAttrs def _getParser(self): op = F8_User._getParser(self) op.add_option("--gecos", type="string") return op ########NEW FILE######## __FILENAME__ = vnc # # Chris Lumens <clumens@redhat.com> # # Copyright 2005, 2006, 2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.errors import * from pykickstart.options import * class FC3_Vnc(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.op = self._getParser() self.enabled = kwargs.get("enabled", False) self.password = kwargs.get("password", "") self.connect = kwargs.get("connect", "") def __str__(self): retval = KickstartCommand.__str__(self) if not self.enabled: return retval retval += "vnc" if self.connect != "": retval += " --connect=%s" % self.connect if self.password != "": retval += " --password=%s" % self.password return retval + "\n" def _getParser(self): op = KSOptionParser() op.add_option("--connect") op.add_option("--password", dest="password") return op def parse(self, args): self.enabled = True (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) self._setToSelf(self.op, opts) return self class FC6_Vnc(FC3_Vnc): removedKeywords = FC3_Vnc.removedKeywords + ["connect"] removedAttrs = FC3_Vnc.removedAttrs + ["connect"] def __init__(self, writePriority=0, host="", port="", *args, **kwargs): FC3_Vnc.__init__(self, writePriority, *args, **kwargs) self.deleteRemovedAttrs() self.host = kwargs.get("host", "") self.port = kwargs.get("port", "") def __str__(self): retval = KickstartCommand.__str__(self) if not self.enabled: return retval retval += "vnc" if self.host != "": retval += " --host=%s" % self.host if self.port != "": retval += " --port=%s" % self.port if self.password != "": retval += " --password=%s" % self.password return retval + "\n" def _getParser(self): def connect_cb (option, opt_str, value, parser): cargs = value.split(":") parser.values.ensure_value("host", cargs[0]) if len(cargs) > 1: parser.values.ensure_value("port", cargs[1]) op = FC3_Vnc._getParser(self) op.add_option("--connect", action="callback", callback=connect_cb, nargs=1, type="string") op.add_option("--host", dest="host") op.add_option("--port", dest="port") return op class F9_Vnc(FC6_Vnc): removedKeywords = FC6_Vnc.removedKeywords removedAttrs = FC6_Vnc.removedAttrs def _getParser(self): op = FC6_Vnc._getParser(self) op.remove_option("--connect") return op ########NEW FILE######## __FILENAME__ = volgroup # # Chris Lumens <clumens@redhat.com> # # Copyright 2005, 2006, 2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.options import * import gettext import warnings _ = lambda x: gettext.ldgettext("pykickstart", x) class FC3_VolGroupData(BaseData): removedKeywords = BaseData.removedKeywords removedAttrs = BaseData.removedAttrs def __init__(self, *args, **kwargs): BaseData.__init__(self, *args, **kwargs) self.format = kwargs.get("format", True) self.pesize = kwargs.get("pesize", 32768) self.preexist = kwargs.get("preexist", False) self.vgname = kwargs.get("vgname", "") self.physvols = kwargs.get("physvols", []) def __eq__(self, y): return self.vgname == y.vgname def __str__(self): retval = BaseData.__str__(self) retval += "volgroup %s" % self.vgname if not self.format: retval += " --noformat" if self.pesize != 0: retval += " --pesize=%d" % self.pesize if self.preexist: retval += " --useexisting" return retval + " " + " ".join(self.physvols) + "\n" class FC3_VolGroup(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=132, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.op = self._getParser() self.vgList = kwargs.get("vgList", []) def __str__(self): retval = "" for vg in self.vgList: retval += vg.__str__() return retval def _getParser(self): # Have to be a little more complicated to set two values. def vg_cb (option, opt_str, value, parser): parser.values.format = False parser.values.preexist = True op = KSOptionParser() op.add_option("--noformat", action="callback", callback=vg_cb, dest="format", default=True, nargs=0) op.add_option("--pesize", dest="pesize", type="int", nargs=1, default=32768) op.add_option("--useexisting", dest="preexist", action="store_true", default=False) return op def parse(self, args): (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) vg = self.handler.VolGroupData() self._setToObj(self.op, opts, vg) vg.lineno = self.lineno vg.vgname = extra[0] vg.physvols = extra[1:] # Check for duplicates in the data list. if vg in self.dataList(): warnings.warn(_("A volgroup with the name %s has already been defined.") % vg.vgname) return vg def dataList(self): return self.vgList ########NEW FILE######## __FILENAME__ = xconfig # # Chris Lumens <clumens@redhat.com> # # Copyright 2005, 2006, 2007, 2008 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.errors import * from pykickstart.options import * import gettext _ = lambda x: gettext.ldgettext("pykickstart", x) class FC3_XConfig(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=0, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.op = self._getParser() self.card = kwargs.get("card", "") self.defaultdesktop = kwargs.get("defaultdesktop", "") self.depth = kwargs.get("depth", 0) self.hsync = kwargs.get("hsync", "") self.monitor = kwargs.get("monitor", "") self.noProbe = kwargs.get("noProbe", False) self.resolution = kwargs.get("resolution", "") self.server = kwargs.get("server", "") self.startX = kwargs.get("startX", False) self.videoRam = kwargs.get("videoRam", "") self.vsync = kwargs.get("vsync", "") def __str__(self): retval = KickstartCommand.__str__(self) if self.card != "": retval += " --card=%s" % self.card if self.defaultdesktop != "": retval += " --defaultdesktop=%s" % self.defaultdesktop if self.depth != 0: retval += " --depth=%d" % self.depth if self.hsync != "": retval += " --hsync=%s" % self.hsync if self.monitor != "": retval += " --monitor=%s" % self.monitor if self.noProbe: retval += " --noprobe" if self.resolution != "": retval += " --resolution=%s" % self.resolution if self.server != "": retval += " --server=%s" % self.server if self.startX: retval += " --startxonboot" if self.videoRam != "": retval += " --videoram=%s" % self.videoRam if self.vsync != "": retval += " --vsync=%s" % self.vsync if retval != "": retval = "# X Window System configuration information\nxconfig %s\n" % retval return retval def _getParser(self): op = KSOptionParser() op.add_option("--card") op.add_option("--defaultdesktop") op.add_option("--depth", action="store", type="int", nargs=1) op.add_option("--hsync") op.add_option("--monitor") op.add_option("--noprobe", dest="noProbe", action="store_true", default=False) op.add_option("--resolution") op.add_option("--server") op.add_option("--startxonboot", dest="startX", action="store_true", default=False) op.add_option("--videoram", dest="videoRam") op.add_option("--vsync") return op def parse(self, args): (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) if extra: mapping = {"command": "xconfig", "options": extra} raise KickstartValueError, formatErrorMsg(self.lineno, msg=_("Unexpected arguments to %(command)s command: %(options)s") % mapping) self._setToSelf(self.op, opts) return self class FC6_XConfig(FC3_XConfig): removedKeywords = FC3_XConfig.removedKeywords + ["card", "hsync", "monitor", "noProbe", "vsync"] removedAttrs = FC3_XConfig.removedAttrs + ["card", "hsync", "monitor", "noProbe", "vsync"] def __init__(self, writePriority=0, *args, **kwargs): FC3_XConfig.__init__(self, writePriority, *args, **kwargs) self.deleteRemovedAttrs() self.driver = kwargs.get("driver", "") def __str__(self): retval = KickstartCommand.__str__(self) if hasattr(self, "driver") and self.driver != "": retval += " --driver=%s" % self.driver if self.defaultdesktop != "": retval += " --defaultdesktop=%s" % self.defaultdesktop if self.depth != 0: retval += " --depth=%d" % self.depth if hasattr(self, "resolution") and self.resolution != "": retval += " --resolution=%s" % self.resolution if self.startX: retval += " --startxonboot" if hasattr(self, "videoRam") and self.videoRam != "": retval += " --videoram=%s" % self.videoRam if retval != "": retval = "# X Window System configuration information\nxconfig %s\n" % retval return retval def _getParser(self): op = FC3_XConfig._getParser(self) op.add_option("--card", deprecated=1) op.add_option("--driver", dest="driver") op.add_option("--hsync", deprecated=1) op.add_option("--monitor", deprecated=1) op.add_option("--noprobe", deprecated=1) op.add_option("--vsync", deprecated=1) return op class F9_XConfig(FC6_XConfig): removedKeywords = FC6_XConfig.removedKeywords removedAttrs = FC6_XConfig.removedAttrs def _getParser(self): op = FC6_XConfig._getParser(self) op.remove_option("--card") op.remove_option("--hsync") op.remove_option("--monitor") op.remove_option("--noprobe") op.remove_option("--vsync") return op class F10_XConfig(F9_XConfig): removedKeywords = F9_XConfig.removedKeywords + ["driver", "resolution", "videoRam"] removedAttrs = F9_XConfig.removedAttrs + ["driver", "resolution", "videoRam"] def __init__(self, writePriority=0, *args, **kwargs): F9_XConfig.__init__(self, writePriority, *args, **kwargs) self.deleteRemovedAttrs() def _getParser(self): op = F9_XConfig._getParser(self) op.add_option("--driver", deprecated=1) op.add_option("--depth", deprecated=1) op.add_option("--resolution", deprecated=1) op.add_option("--videoram", deprecated=1) return op class F14_XConfig(F10_XConfig): removedKeywords = F10_XConfig.removedKeywords removedAttrs = F10_XConfig.removedAttrs def _getParser(self): op = F10_XConfig._getParser(self) op.remove_option("--driver") op.remove_option("--depth") op.remove_option("--resolution") op.remove_option("--videoram") return op ########NEW FILE######## __FILENAME__ = zerombr # # Chris Lumens <clumens@redhat.com> # # Copyright 2005, 2006, 2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # import warnings from pykickstart.base import * from pykickstart.options import * import gettext _ = lambda x: gettext.ldgettext("pykickstart", x) class FC3_ZeroMbr(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=110, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.op = self._getParser() self.zerombr = kwargs.get("zerombr", False) def __str__(self): retval = KickstartCommand.__str__(self) if self.zerombr: retval += "# Clear the Master Boot Record\nzerombr\n" return retval def _getParser(self): op = KSOptionParser() return op def parse(self, args): (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) if len(extra) > 0: warnings.warn(_("Ignoring deprecated option on line %s: The zerombr command no longer takes any options. In future releases, this will result in a fatal error from kickstart. Please modify your kickstart file to remove any options.") % self.lineno, DeprecationWarning) self.zerombr = True return self class F9_ZeroMbr(FC3_ZeroMbr): removedKeywords = FC3_ZeroMbr.removedKeywords removedAttrs = FC3_ZeroMbr.removedAttrs def parse(self, args): (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) if len(extra) > 0: raise KickstartParseError, formatErrorMsg(self.lineno, msg=_("Kickstart command %s does not take any arguments") % "zerombr") self.zerombr = True return self ########NEW FILE######## __FILENAME__ = zfcp # # Chris Lumens <clumens@redhat.com> # # Copyright 2005, 2006, 2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.options import * import gettext import warnings _ = lambda x: gettext.ldgettext("pykickstart", x) class FC3_ZFCPData(BaseData): removedKeywords = BaseData.removedKeywords removedAttrs = BaseData.removedAttrs def __init__(self, *args, **kwargs): BaseData.__init__(self, *args, **kwargs) self.devnum = kwargs.get("devnum", "") self.wwpn = kwargs.get("wwpn", "") self.fcplun = kwargs.get("fcplun", "") self.scsiid = kwargs.get("scsiid", "") self.scsilun = kwargs.get("scsilun", "") def __eq__(self, y): return self.devnum == y.devnum and self.wwpn == y.wwpn and \ self.fcplun == y.fcplun and self.scsiid == y.scsiid and \ self.scsilun == y.scsilun def __str__(self): retval = BaseData.__str__(self) retval += "zfcp" if self.devnum != "": retval += " --devnum=%s" % self.devnum if self.wwpn != "": retval += " --wwpn=%s" % self.wwpn if self.fcplun != "": retval += " --fcplun=%s" % self.fcplun if hasattr(self, "scsiid") and self.scsiid != "": retval += " --scsiid=%s" % self.scsiid if hasattr(self, "scsilun") and self.scsilun != "": retval += " --scsilun=%s" % self.scsilun return retval + "\n" class F12_ZFCPData(FC3_ZFCPData): removedKeywords = FC3_ZFCPData.removedKeywords + ["scsiid", "scsilun"] removedAttrs = FC3_ZFCPData.removedAttrs + ["scsiid", "scsilun"] def __init__(self, *args, **kwargs): FC3_ZFCPData.__init__(self, *args, **kwargs) self.deleteRemovedAttrs() F14_ZFCPData = F12_ZFCPData class FC3_ZFCP(KickstartCommand): removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, writePriority=71, *args, **kwargs): KickstartCommand.__init__(self, writePriority, *args, **kwargs) self.op = self._getParser() self.zfcp = kwargs.get("zfcp", []) def __str__(self): retval = "" for zfcp in self.zfcp: retval += zfcp.__str__() return retval def _getParser(self): op = KSOptionParser() op.add_option("--devnum", dest="devnum", required=1) op.add_option("--fcplun", dest="fcplun", required=1) op.add_option("--scsiid", dest="scsiid", required=1) op.add_option("--scsilun", dest="scsilun", required=1) op.add_option("--wwpn", dest="wwpn", required=1) return op def parse(self, args): zd = self.handler.ZFCPData() (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) self._setToObj(self.op, opts, zd) zd.lineno = self.lineno # Check for duplicates in the data list. if zd in self.dataList(): warnings.warn(_("A zfcp with this information has already been defined.")) return zd def dataList(self): return self.zfcp class F12_ZFCP(FC3_ZFCP): removedKeywords = FC3_ZFCP.removedKeywords removedAttrs = FC3_ZFCP.removedAttrs + ["scsiid", "scsilun"] def __init__(self, *args, **kwargs): FC3_ZFCP.__init__(self, *args, **kwargs) self.deleteRemovedAttrs() def _getParser(self): op = FC3_ZFCP._getParser(self) op.add_option("--scsiid", deprecated=1) op.add_option("--scsilun", deprecated=1) return op class F14_ZFCP(F12_ZFCP): removedKeywords = F12_ZFCP.removedKeywords removedAttrs = F12_ZFCP.removedAttrs def _getParser(self): op = F12_ZFCP._getParser(self) op.remove_option("--scsiid") op.remove_option("--scsilun") return op ########NEW FILE######## __FILENAME__ = constants # # Chris Lumens <clumens@redhat.com> # # Copyright 2005-2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # CLEARPART_TYPE_LINUX = 0 CLEARPART_TYPE_ALL = 1 CLEARPART_TYPE_NONE = 2 DISPLAY_MODE_CMDLINE = 0 DISPLAY_MODE_GRAPHICAL = 1 DISPLAY_MODE_TEXT = 2 FIRSTBOOT_DEFAULT = 0 FIRSTBOOT_SKIP = 1 FIRSTBOOT_RECONFIG = 2 KS_MISSING_PROMPT = 0 KS_MISSING_IGNORE = 1 SELINUX_DISABLED = 0 SELINUX_ENFORCING = 1 SELINUX_PERMISSIVE = 2 KS_SCRIPT_PRE = 0 KS_SCRIPT_POST = 1 KS_SCRIPT_TRACEBACK = 2 KS_WAIT = 0 KS_REBOOT = 1 KS_SHUTDOWN = 2 KS_INSTKEY_SKIP = -99 BOOTPROTO_DHCP = "dhcp" BOOTPROTO_BOOTP = "bootp" BOOTPROTO_STATIC = "static" BOOTPROTO_QUERY = "query" BOOTPROTO_IBFT = "ibft" GROUP_REQUIRED = 0 GROUP_DEFAULT = 1 GROUP_ALL = 2 ########NEW FILE######## __FILENAME__ = errors # # errors.py: Kickstart error handling. # # Chris Lumens <clumens@redhat.com> # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # """ Error handling classes and functions. This module exports a single function: formatErrorMsg - Properly formats an error message. It also exports several exception classes: KickstartError - A generic exception class. KickstartParseError - An exception for errors relating to parsing. KickstartValueError - An exception for errors relating to option processing. KickstartVersionError - An exception for errors relating to unsupported syntax versions. """ import gettext _ = lambda x: gettext.ldgettext("pykickstart", x) def formatErrorMsg(lineno, msg=""): """Properly format the error message msg for inclusion in an exception.""" if msg != "": mapping = {"lineno": lineno, "msg": msg} return _("The following problem occurred on line %(lineno)s of the kickstart file:\n\n%(msg)s\n") % mapping else: return _("There was a problem reading from line %s of the kickstart file") % lineno class KickstartError(Exception): """A generic exception class for unspecific error conditions.""" def __init__(self, val = ""): """Create a new KickstartError exception instance with the descriptive message val. val should be the return value of formatErrorMsg. """ Exception.__init__(self) self.value = val def __str__ (self): return self.value class KickstartParseError(KickstartError): """An exception class for errors when processing the input file, such as unknown options, commands, or sections. """ def __init__(self, msg): """Create a new KickstartParseError exception instance with the descriptive message val. val should be the return value of formatErrorMsg. """ KickstartError.__init__(self, msg) def __str__(self): return self.value class KickstartValueError(KickstartError): """An exception class for errors when processing arguments to commands, such as too many arguments, too few arguments, or missing required arguments. """ def __init__(self, msg): """Create a new KickstartValueError exception instance with the descriptive message val. val should be the return value of formatErrorMsg. """ KickstartError.__init__(self, msg) def __str__ (self): return self.value class KickstartVersionError(KickstartError): """An exception class for errors related to using an incorrect version of kickstart syntax. """ def __init__(self, msg): """Create a new KickstartVersionError exception instance with the descriptive message val. val should be the return value of formatErrorMsg. """ KickstartError.__init__(self, msg) def __str__ (self): return self.value ########NEW FILE######## __FILENAME__ = control # # Chris Lumens <clumens@redhat.com> # # Copyright 2007, 2008, 2009, 2010 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.version import * from pykickstart.commands import * # This map is keyed on kickstart syntax version as provided by # pykickstart.version. Within each sub-dict is a mapping from command name # to the class that handles it. This is an onto mapping - that is, multiple # command names can map to the same class. However, the Handler will ensure # that only one instance of each class ever exists. commandMap = { FC3: { "auth": authconfig.FC3_Authconfig, "authconfig": authconfig.FC3_Authconfig, "autopart": autopart.FC3_AutoPart, "autostep": autostep.FC3_AutoStep, "bootloader": bootloader.FC3_Bootloader, "cdrom": method.FC3_Method, "clearpart": clearpart.FC3_ClearPart, "cmdline": displaymode.FC3_DisplayMode, "device": device.FC3_Device, "deviceprobe": deviceprobe.FC3_DeviceProbe, "driverdisk": driverdisk.FC3_DriverDisk, "firewall": firewall.FC3_Firewall, "firstboot": firstboot.FC3_Firstboot, "graphical": displaymode.FC3_DisplayMode, "halt": reboot.FC3_Reboot, "harddrive": method.FC3_Method, "ignoredisk": ignoredisk.FC3_IgnoreDisk, "install": upgrade.FC3_Upgrade, "interactive": interactive.FC3_Interactive, "keyboard": keyboard.FC3_Keyboard, "lang": lang.FC3_Lang, "langsupport": langsupport.FC3_LangSupport, "lilo": bootloader.FC3_Bootloader, "lilocheck": lilocheck.FC3_LiloCheck, "logvol": logvol.FC3_LogVol, "monitor": monitor.FC3_Monitor, "mouse": mouse.FC3_Mouse, "network": network.FC3_Network, "nfs": method.FC3_Method, "part": partition.FC3_Partition, "partition": partition.FC3_Partition, "poweroff": reboot.FC3_Reboot, "raid": raid.FC3_Raid, "reboot": reboot.FC3_Reboot, "rootpw": rootpw.FC3_RootPw, "selinux": selinux.FC3_SELinux, "shutdown": reboot.FC3_Reboot, "skipx": skipx.FC3_SkipX, "text": displaymode.FC3_DisplayMode, "timezone": timezone.FC3_Timezone, "upgrade": upgrade.FC3_Upgrade, "url": method.FC3_Method, "vnc": vnc.FC3_Vnc, "volgroup": volgroup.FC3_VolGroup, "xconfig": xconfig.FC3_XConfig, "zerombr": zerombr.FC3_ZeroMbr, "zfcp": zfcp.FC3_ZFCP, }, # based on fc3 FC4: { "auth": authconfig.FC3_Authconfig, "authconfig": authconfig.FC3_Authconfig, "autopart": autopart.FC3_AutoPart, "autostep": autostep.FC3_AutoStep, "bootloader": bootloader.FC4_Bootloader, "cdrom": method.FC3_Method, "clearpart": clearpart.FC3_ClearPart, "cmdline": displaymode.FC3_DisplayMode, "device": device.FC3_Device, "deviceprobe": deviceprobe.FC3_DeviceProbe, "driverdisk": driverdisk.FC4_DriverDisk, "firewall": firewall.FC3_Firewall, "firstboot": firstboot.FC3_Firstboot, "graphical": displaymode.FC3_DisplayMode, "halt": reboot.FC3_Reboot, "harddrive": method.FC3_Method, "ignoredisk": ignoredisk.FC3_IgnoreDisk, "install": upgrade.FC3_Upgrade, "interactive": interactive.FC3_Interactive, "keyboard": keyboard.FC3_Keyboard, "lang": lang.FC3_Lang, "langsupport": langsupport.FC3_LangSupport, "logvol": logvol.FC4_LogVol, "mediacheck": mediacheck.FC4_MediaCheck, "monitor": monitor.FC3_Monitor, "mouse": mouse.FC3_Mouse, "network": network.FC4_Network, "nfs": method.FC3_Method, "part": partition.FC4_Partition, "partition": partition.FC4_Partition, "poweroff": reboot.FC3_Reboot, "raid": raid.FC4_Raid, "reboot": reboot.FC3_Reboot, "rootpw": rootpw.FC3_RootPw, "selinux": selinux.FC3_SELinux, "shutdown": reboot.FC3_Reboot, "skipx": skipx.FC3_SkipX, "text": displaymode.FC3_DisplayMode, "timezone": timezone.FC3_Timezone, "upgrade": upgrade.FC3_Upgrade, "url": method.FC3_Method, "vnc": vnc.FC3_Vnc, "volgroup": volgroup.FC3_VolGroup, "xconfig": xconfig.FC3_XConfig, "zerombr": zerombr.FC3_ZeroMbr, "zfcp": zfcp.FC3_ZFCP, }, # based on fc4 FC5: { "auth": authconfig.FC3_Authconfig, "authconfig": authconfig.FC3_Authconfig, "autopart": autopart.FC3_AutoPart, "autostep": autostep.FC3_AutoStep, "bootloader": bootloader.FC4_Bootloader, "cdrom": method.FC3_Method, "clearpart": clearpart.FC3_ClearPart, "cmdline": displaymode.FC3_DisplayMode, "device": device.FC3_Device, "deviceprobe": deviceprobe.FC3_DeviceProbe, "driverdisk": driverdisk.FC4_DriverDisk, "firewall": firewall.FC3_Firewall, "firstboot": firstboot.FC3_Firstboot, "graphical": displaymode.FC3_DisplayMode, "halt": reboot.FC3_Reboot, "harddrive": method.FC3_Method, "ignoredisk": ignoredisk.FC3_IgnoreDisk, "install": upgrade.FC3_Upgrade, "interactive": interactive.FC3_Interactive, "keyboard": keyboard.FC3_Keyboard, "lang": lang.FC3_Lang, "langsupport": langsupport.FC5_LangSupport, "logvol": logvol.FC4_LogVol, "mediacheck": mediacheck.FC4_MediaCheck, "monitor": monitor.FC3_Monitor, "mouse": mouse.FC3_Mouse, "network": network.FC4_Network, "nfs": method.FC3_Method, "part": partition.FC4_Partition, "partition": partition.FC4_Partition, "poweroff": reboot.FC3_Reboot, "raid": raid.FC5_Raid, "reboot": reboot.FC3_Reboot, "rootpw": rootpw.FC3_RootPw, "selinux": selinux.FC3_SELinux, "shutdown": reboot.FC3_Reboot, "skipx": skipx.FC3_SkipX, "text": displaymode.FC3_DisplayMode, "timezone": timezone.FC3_Timezone, "upgrade": upgrade.FC3_Upgrade, "url": method.FC3_Method, "vnc": vnc.FC3_Vnc, "volgroup": volgroup.FC3_VolGroup, "xconfig": xconfig.FC3_XConfig, "zerombr": zerombr.FC3_ZeroMbr, "zfcp": zfcp.FC3_ZFCP, }, # based on fc5 FC6: { "auth": authconfig.FC3_Authconfig, "authconfig": authconfig.FC3_Authconfig, "autopart": autopart.FC3_AutoPart, "autostep": autostep.FC3_AutoStep, "bootloader": bootloader.FC4_Bootloader, "cdrom": method.FC6_Method, "clearpart": clearpart.FC3_ClearPart, "cmdline": displaymode.FC3_DisplayMode, "device": device.FC3_Device, "deviceprobe": deviceprobe.FC3_DeviceProbe, "dmraid": dmraid.FC6_DmRaid, "driverdisk": driverdisk.FC4_DriverDisk, "firewall": firewall.FC3_Firewall, "firstboot": firstboot.FC3_Firstboot, "graphical": displaymode.FC3_DisplayMode, "halt": reboot.FC6_Reboot, "harddrive": method.FC6_Method, "ignoredisk": ignoredisk.FC3_IgnoreDisk, "install": upgrade.FC3_Upgrade, "interactive": interactive.FC3_Interactive, "iscsi": iscsi.FC6_Iscsi, "iscsiname": iscsiname.FC6_IscsiName, "keyboard": keyboard.FC3_Keyboard, "lang": lang.FC3_Lang, "langsupport": langsupport.FC5_LangSupport, "logging": logging.FC6_Logging, "logvol": logvol.FC4_LogVol, "mediacheck": mediacheck.FC4_MediaCheck, "monitor": monitor.FC6_Monitor, "mouse": mouse.FC3_Mouse, "multipath": multipath.FC6_MultiPath, "network": network.FC6_Network, "nfs": method.FC6_Method, "part": partition.FC4_Partition, "partition": partition.FC4_Partition, "poweroff": reboot.FC6_Reboot, "raid": raid.FC5_Raid, "reboot": reboot.FC6_Reboot, "repo": repo.FC6_Repo, "rootpw": rootpw.FC3_RootPw, "selinux": selinux.FC3_SELinux, "services": services.FC6_Services, "shutdown": reboot.FC6_Reboot, "skipx": skipx.FC3_SkipX, "text": displaymode.FC3_DisplayMode, "timezone": timezone.FC6_Timezone, "upgrade": upgrade.FC3_Upgrade, "user": user.FC6_User, "url": method.FC6_Method, "vnc": vnc.FC6_Vnc, "volgroup": volgroup.FC3_VolGroup, "xconfig": xconfig.FC6_XConfig, "zerombr": zerombr.FC3_ZeroMbr, "zfcp": zfcp.FC3_ZFCP, }, # based on fc6 F7: { "auth": authconfig.FC3_Authconfig, "authconfig": authconfig.FC3_Authconfig, "autopart": autopart.FC3_AutoPart, "autostep": autostep.FC3_AutoStep, "bootloader": bootloader.FC4_Bootloader, "cdrom": method.FC6_Method, "clearpart": clearpart.FC3_ClearPart, "cmdline": displaymode.FC3_DisplayMode, "device": device.FC3_Device, "deviceprobe": deviceprobe.FC3_DeviceProbe, "dmraid": dmraid.FC6_DmRaid, "driverdisk": driverdisk.FC4_DriverDisk, "firewall": firewall.FC3_Firewall, "firstboot": firstboot.FC3_Firstboot, "graphical": displaymode.FC3_DisplayMode, "halt": reboot.FC6_Reboot, "harddrive": method.FC6_Method, "ignoredisk": ignoredisk.FC3_IgnoreDisk, "install": upgrade.FC3_Upgrade, "interactive": interactive.FC3_Interactive, "iscsi": iscsi.FC6_Iscsi, "iscsiname": iscsiname.FC6_IscsiName, "keyboard": keyboard.FC3_Keyboard, "lang": lang.FC3_Lang, "logging": logging.FC6_Logging, "logvol": logvol.FC4_LogVol, "mediacheck": mediacheck.FC4_MediaCheck, "monitor": monitor.FC6_Monitor, "multipath": multipath.FC6_MultiPath, "network": network.FC6_Network, "nfs": method.FC6_Method, "part": partition.FC4_Partition, "partition": partition.FC4_Partition, "poweroff": reboot.FC6_Reboot, "raid": raid.F7_Raid, "reboot": reboot.FC6_Reboot, "repo": repo.FC6_Repo, "rootpw": rootpw.FC3_RootPw, "selinux": selinux.FC3_SELinux, "services": services.FC6_Services, "shutdown": reboot.FC6_Reboot, "skipx": skipx.FC3_SkipX, "text": displaymode.FC3_DisplayMode, "timezone": timezone.FC6_Timezone, "updates": updates.F7_Updates, "upgrade": upgrade.FC3_Upgrade, "url": method.FC6_Method, "user": user.FC6_User, "vnc": vnc.FC6_Vnc, "volgroup": volgroup.FC3_VolGroup, "xconfig": xconfig.FC6_XConfig, "zerombr": zerombr.FC3_ZeroMbr, "zfcp": zfcp.FC3_ZFCP, }, # based on f7 F8: { "auth": authconfig.FC3_Authconfig, "authconfig": authconfig.FC3_Authconfig, "autopart": autopart.FC3_AutoPart, "autostep": autostep.FC3_AutoStep, "bootloader": bootloader.F8_Bootloader, "cdrom": method.FC6_Method, "clearpart": clearpart.FC3_ClearPart, "cmdline": displaymode.FC3_DisplayMode, "device": device.F8_Device, "deviceprobe": deviceprobe.FC3_DeviceProbe, "dmraid": dmraid.FC6_DmRaid, "driverdisk": driverdisk.FC4_DriverDisk, "firewall": firewall.FC3_Firewall, "firstboot": firstboot.FC3_Firstboot, "graphical": displaymode.FC3_DisplayMode, "halt": reboot.FC6_Reboot, "harddrive": method.FC6_Method, "ignoredisk": ignoredisk.F8_IgnoreDisk, "install": upgrade.FC3_Upgrade, "interactive": interactive.FC3_Interactive, "iscsi": iscsi.FC6_Iscsi, "iscsiname": iscsiname.FC6_IscsiName, "keyboard": keyboard.FC3_Keyboard, "lang": lang.FC3_Lang, "logging": logging.FC6_Logging, "logvol": logvol.FC4_LogVol, "mediacheck": mediacheck.FC4_MediaCheck, "monitor": monitor.FC6_Monitor, "multipath": multipath.FC6_MultiPath, "network": network.F8_Network, "nfs": method.FC6_Method, "part": partition.FC4_Partition, "partition": partition.FC4_Partition, "poweroff": reboot.FC6_Reboot, "raid": raid.F7_Raid, "reboot": reboot.FC6_Reboot, "repo": repo.F8_Repo, "rootpw": rootpw.F8_RootPw, "selinux": selinux.FC3_SELinux, "services": services.FC6_Services, "shutdown": reboot.FC6_Reboot, "skipx": skipx.FC3_SkipX, "text": displaymode.FC3_DisplayMode, "timezone": timezone.FC6_Timezone, "updates": updates.F7_Updates, "upgrade": upgrade.FC3_Upgrade, "url": method.FC6_Method, "user": user.F8_User, "vnc": vnc.FC6_Vnc, "volgroup": volgroup.FC3_VolGroup, "xconfig": xconfig.FC6_XConfig, "zerombr": zerombr.FC3_ZeroMbr, "zfcp": zfcp.FC3_ZFCP, }, # based on f8 F9: { "auth": authconfig.FC3_Authconfig, "authconfig": authconfig.FC3_Authconfig, "autopart": autopart.F9_AutoPart, "autostep": autostep.FC3_AutoStep, "bootloader": bootloader.F8_Bootloader, "cdrom": method.FC6_Method, "clearpart": clearpart.FC3_ClearPart, "cmdline": displaymode.FC3_DisplayMode, "device": device.F8_Device, "deviceprobe": deviceprobe.FC3_DeviceProbe, "dmraid": dmraid.FC6_DmRaid, "driverdisk": driverdisk.FC4_DriverDisk, "firewall": firewall.F9_Firewall, "firstboot": firstboot.FC3_Firstboot, "graphical": displaymode.FC3_DisplayMode, "halt": reboot.FC6_Reboot, "harddrive": method.FC6_Method, "ignoredisk": ignoredisk.F8_IgnoreDisk, "install": upgrade.FC3_Upgrade, "interactive": interactive.FC3_Interactive, "iscsi": iscsi.FC6_Iscsi, "iscsiname": iscsiname.FC6_IscsiName, "keyboard": keyboard.FC3_Keyboard, "lang": lang.FC3_Lang, "logging": logging.FC6_Logging, "logvol": logvol.F9_LogVol, "mediacheck": mediacheck.FC4_MediaCheck, "monitor": monitor.FC6_Monitor, "multipath": multipath.FC6_MultiPath, "network": network.F9_Network, "nfs": method.FC6_Method, "part": partition.F9_Partition, "partition": partition.F9_Partition, "poweroff": reboot.FC6_Reboot, "raid": raid.F9_Raid, "reboot": reboot.FC6_Reboot, "repo": repo.F8_Repo, "rootpw": rootpw.F8_RootPw, "selinux": selinux.FC3_SELinux, "services": services.FC6_Services, "shutdown": reboot.FC6_Reboot, "skipx": skipx.FC3_SkipX, "text": displaymode.FC3_DisplayMode, "timezone": timezone.FC6_Timezone, "updates": updates.F7_Updates, "upgrade": upgrade.FC3_Upgrade, "url": method.FC6_Method, "user": user.F8_User, "vnc": vnc.F9_Vnc, "volgroup": volgroup.FC3_VolGroup, "xconfig": xconfig.F9_XConfig, "zerombr": zerombr.F9_ZeroMbr, "zfcp": zfcp.FC3_ZFCP, }, # based on f9 F10: { "auth": authconfig.FC3_Authconfig, "authconfig": authconfig.FC3_Authconfig, "autopart": autopart.F9_AutoPart, "autostep": autostep.FC3_AutoStep, "bootloader": bootloader.F8_Bootloader, "cdrom": method.FC6_Method, "clearpart": clearpart.FC3_ClearPart, "cmdline": displaymode.FC3_DisplayMode, "device": device.F8_Device, "deviceprobe": deviceprobe.FC3_DeviceProbe, "dmraid": dmraid.FC6_DmRaid, "driverdisk": driverdisk.FC4_DriverDisk, "firewall": firewall.F10_Firewall, "firstboot": firstboot.FC3_Firstboot, "graphical": displaymode.FC3_DisplayMode, "halt": reboot.FC6_Reboot, "harddrive": method.FC6_Method, "ignoredisk": ignoredisk.F8_IgnoreDisk, "install": upgrade.FC3_Upgrade, "interactive": interactive.FC3_Interactive, "iscsi": iscsi.F10_Iscsi, "iscsiname": iscsiname.FC6_IscsiName, "keyboard": keyboard.FC3_Keyboard, "lang": lang.FC3_Lang, "logging": logging.FC6_Logging, "logvol": logvol.F9_LogVol, "mediacheck": mediacheck.FC4_MediaCheck, "monitor": monitor.F10_Monitor, "multipath": multipath.FC6_MultiPath, "network": network.F9_Network, "nfs": method.FC6_Method, "part": partition.F9_Partition, "partition": partition.F9_Partition, "poweroff": reboot.FC6_Reboot, "raid": raid.F9_Raid, "reboot": reboot.FC6_Reboot, "repo": repo.F8_Repo, "rescue": rescue.F10_Rescue, "rootpw": rootpw.F8_RootPw, "selinux": selinux.FC3_SELinux, "services": services.FC6_Services, "shutdown": reboot.FC6_Reboot, "skipx": skipx.FC3_SkipX, "text": displaymode.FC3_DisplayMode, "timezone": timezone.FC6_Timezone, "updates": updates.F7_Updates, "upgrade": upgrade.FC3_Upgrade, "url": method.FC6_Method, "user": user.F8_User, "vnc": vnc.F9_Vnc, "volgroup": volgroup.FC3_VolGroup, "xconfig": xconfig.F10_XConfig, "zerombr": zerombr.F9_ZeroMbr, "zfcp": zfcp.FC3_ZFCP, }, # based on f10 F11: { "auth": authconfig.FC3_Authconfig, "authconfig": authconfig.FC3_Authconfig, "autopart": autopart.F9_AutoPart, "autostep": autostep.FC3_AutoStep, "bootloader": bootloader.F8_Bootloader, "cdrom": method.FC6_Method, "clearpart": clearpart.FC3_ClearPart, "cmdline": displaymode.FC3_DisplayMode, "device": device.F8_Device, "deviceprobe": deviceprobe.FC3_DeviceProbe, "dmraid": dmraid.FC6_DmRaid, "driverdisk": driverdisk.FC4_DriverDisk, "firewall": firewall.F10_Firewall, "firstboot": firstboot.FC3_Firstboot, "graphical": displaymode.FC3_DisplayMode, "halt": reboot.FC6_Reboot, "harddrive": method.FC6_Method, "ignoredisk": ignoredisk.F8_IgnoreDisk, "install": upgrade.F11_Upgrade, "interactive": interactive.FC3_Interactive, "iscsi": iscsi.F10_Iscsi, "iscsiname": iscsiname.FC6_IscsiName, "keyboard": keyboard.FC3_Keyboard, "lang": lang.FC3_Lang, "logging": logging.FC6_Logging, "logvol": logvol.F9_LogVol, "mediacheck": mediacheck.FC4_MediaCheck, "monitor": monitor.F10_Monitor, "multipath": multipath.FC6_MultiPath, "network": network.F9_Network, "nfs": method.FC6_Method, "part": partition.F11_Partition, "partition": partition.F11_Partition, "poweroff": reboot.FC6_Reboot, "raid": raid.F9_Raid, "reboot": reboot.FC6_Reboot, "repo": repo.F11_Repo, "rescue": rescue.F10_Rescue, "rootpw": rootpw.F8_RootPw, "selinux": selinux.FC3_SELinux, "services": services.FC6_Services, "shutdown": reboot.FC6_Reboot, "skipx": skipx.FC3_SkipX, "text": displaymode.FC3_DisplayMode, "timezone": timezone.FC6_Timezone, "updates": updates.F7_Updates, "upgrade": upgrade.F11_Upgrade, "url": method.FC6_Method, "user": user.F8_User, "vnc": vnc.F9_Vnc, "volgroup": volgroup.FC3_VolGroup, "xconfig": xconfig.F10_XConfig, "zerombr": zerombr.F9_ZeroMbr, "zfcp": zfcp.FC3_ZFCP, }, # based on f11 F12: { "auth": authconfig.FC3_Authconfig, "authconfig": authconfig.FC3_Authconfig, "autopart": autopart.F12_AutoPart, "autostep": autostep.FC3_AutoStep, "bootloader": bootloader.F12_Bootloader, "cdrom": method.FC6_Method, "clearpart": clearpart.FC3_ClearPart, "cmdline": displaymode.FC3_DisplayMode, "device": device.F8_Device, "deviceprobe": deviceprobe.FC3_DeviceProbe, "dmraid": dmraid.FC6_DmRaid, "driverdisk": driverdisk.F12_DriverDisk, "fcoe": fcoe.F12_Fcoe, "firewall": firewall.F10_Firewall, "firstboot": firstboot.FC3_Firstboot, "graphical": displaymode.FC3_DisplayMode, "group": group.F12_Group, "halt": reboot.FC6_Reboot, "harddrive": method.FC6_Method, "ignoredisk": ignoredisk.F8_IgnoreDisk, "install": upgrade.F11_Upgrade, "interactive": interactive.FC3_Interactive, "iscsi": iscsi.F10_Iscsi, "iscsiname": iscsiname.FC6_IscsiName, "keyboard": keyboard.FC3_Keyboard, "lang": lang.FC3_Lang, "logging": logging.FC6_Logging, "logvol": logvol.F12_LogVol, "mediacheck": mediacheck.FC4_MediaCheck, "monitor": monitor.F10_Monitor, "multipath": multipath.FC6_MultiPath, "network": network.F9_Network, "nfs": method.FC6_Method, "part": partition.F12_Partition, "partition": partition.F12_Partition, "poweroff": reboot.FC6_Reboot, "raid": raid.F12_Raid, "reboot": reboot.FC6_Reboot, "repo": repo.F11_Repo, "rescue": rescue.F10_Rescue, "rootpw": rootpw.F8_RootPw, "selinux": selinux.FC3_SELinux, "services": services.FC6_Services, "shutdown": reboot.FC6_Reboot, "skipx": skipx.FC3_SkipX, "text": displaymode.FC3_DisplayMode, "timezone": timezone.FC6_Timezone, "updates": updates.F7_Updates, "upgrade": upgrade.F11_Upgrade, "url": method.FC6_Method, "user": user.F12_User, "vnc": vnc.F9_Vnc, "volgroup": volgroup.FC3_VolGroup, "xconfig": xconfig.F10_XConfig, "zerombr": zerombr.F9_ZeroMbr, "zfcp": zfcp.F12_ZFCP, }, # based on f12 F13: { "auth": authconfig.FC3_Authconfig, "authconfig": authconfig.FC3_Authconfig, "autopart": autopart.F12_AutoPart, "autostep": autostep.FC3_AutoStep, "bootloader": bootloader.F12_Bootloader, "cdrom": method.F13_Method, "clearpart": clearpart.FC3_ClearPart, "cmdline": displaymode.FC3_DisplayMode, "device": device.F8_Device, "deviceprobe": deviceprobe.FC3_DeviceProbe, "dmraid": dmraid.FC6_DmRaid, "driverdisk": driverdisk.F12_DriverDisk, "fcoe": fcoe.F13_Fcoe, "firewall": firewall.F10_Firewall, "firstboot": firstboot.FC3_Firstboot, "graphical": displaymode.FC3_DisplayMode, "group": group.F12_Group, "halt": reboot.FC6_Reboot, "harddrive": method.F13_Method, "ignoredisk": ignoredisk.F8_IgnoreDisk, "install": upgrade.F11_Upgrade, "interactive": interactive.FC3_Interactive, "iscsi": iscsi.F10_Iscsi, "iscsiname": iscsiname.FC6_IscsiName, "keyboard": keyboard.FC3_Keyboard, "lang": lang.FC3_Lang, "logging": logging.FC6_Logging, "logvol": logvol.F12_LogVol, "mediacheck": mediacheck.FC4_MediaCheck, "monitor": monitor.F10_Monitor, "multipath": multipath.FC6_MultiPath, "network": network.F9_Network, "nfs": method.F13_Method, "part": partition.F12_Partition, "partition": partition.F12_Partition, "poweroff": reboot.FC6_Reboot, "raid": raid.F13_Raid, "reboot": reboot.FC6_Reboot, "repo": repo.F13_Repo, "rescue": rescue.F10_Rescue, "rootpw": rootpw.F8_RootPw, "selinux": selinux.FC3_SELinux, "services": services.FC6_Services, "shutdown": reboot.FC6_Reboot, "skipx": skipx.FC3_SkipX, "sshpw": sshpw.F13_SshPw, "text": displaymode.FC3_DisplayMode, "timezone": timezone.FC6_Timezone, "updates": updates.F7_Updates, "upgrade": upgrade.F11_Upgrade, "url": method.F13_Method, "user": user.F12_User, "vnc": vnc.F9_Vnc, "volgroup": volgroup.FC3_VolGroup, "xconfig": xconfig.F10_XConfig, "zerombr": zerombr.F9_ZeroMbr, "zfcp": zfcp.F12_ZFCP, }, # based on f13 F14: { "auth": authconfig.FC3_Authconfig, "authconfig": authconfig.FC3_Authconfig, "autopart": autopart.F12_AutoPart, "autostep": autostep.FC3_AutoStep, "bootloader": bootloader.F14_Bootloader, "cdrom": method.F14_Method, "clearpart": clearpart.FC3_ClearPart, "cmdline": displaymode.FC3_DisplayMode, "device": device.F8_Device, "deviceprobe": deviceprobe.FC3_DeviceProbe, "dmraid": dmraid.FC6_DmRaid, "driverdisk": driverdisk.F14_DriverDisk, "fcoe": fcoe.F13_Fcoe, "firewall": firewall.F14_Firewall, "firstboot": firstboot.FC3_Firstboot, "graphical": displaymode.FC3_DisplayMode, "group": group.F12_Group, "halt": reboot.FC6_Reboot, "harddrive": method.F14_Method, "ignoredisk": ignoredisk.F14_IgnoreDisk, "install": upgrade.F11_Upgrade, "interactive": interactive.F14_Interactive, "iscsi": iscsi.F10_Iscsi, "iscsiname": iscsiname.FC6_IscsiName, "keyboard": keyboard.FC3_Keyboard, "lang": lang.FC3_Lang, "logging": logging.FC6_Logging, "logvol": logvol.F14_LogVol, "mediacheck": mediacheck.FC4_MediaCheck, "monitor": monitor.F10_Monitor, "multipath": multipath.FC6_MultiPath, "network": network.F9_Network, "nfs": method.F14_Method, "part": partition.F14_Partition, "partition": partition.F14_Partition, "poweroff": reboot.FC6_Reboot, "raid": raid.F14_Raid, "reboot": reboot.FC6_Reboot, "repo": repo.F14_Repo, "rescue": rescue.F10_Rescue, "rootpw": rootpw.F8_RootPw, "selinux": selinux.FC3_SELinux, "services": services.FC6_Services, "shutdown": reboot.FC6_Reboot, "skipx": skipx.FC3_SkipX, "sshpw": sshpw.F13_SshPw, "text": displaymode.FC3_DisplayMode, "timezone": timezone.FC6_Timezone, "updates": updates.F7_Updates, "upgrade": upgrade.F11_Upgrade, "url": method.F14_Method, "user": user.F12_User, "vnc": vnc.F9_Vnc, "volgroup": volgroup.FC3_VolGroup, "xconfig": xconfig.F14_XConfig, "zerombr": zerombr.F9_ZeroMbr, "zfcp": zfcp.F14_ZFCP, }, # based on f14 F15: { "auth": authconfig.FC3_Authconfig, "authconfig": authconfig.FC3_Authconfig, "autopart": autopart.F12_AutoPart, "autostep": autostep.FC3_AutoStep, "bootloader": bootloader.F15_Bootloader, "cdrom": method.F14_Method, "clearpart": clearpart.FC3_ClearPart, "cmdline": displaymode.FC3_DisplayMode, "device": device.F8_Device, "deviceprobe": deviceprobe.FC3_DeviceProbe, "dmraid": dmraid.FC6_DmRaid, "driverdisk": driverdisk.F14_DriverDisk, "fcoe": fcoe.F13_Fcoe, "firewall": firewall.F14_Firewall, "firstboot": firstboot.FC3_Firstboot, "graphical": displaymode.FC3_DisplayMode, "group": group.F12_Group, "halt": reboot.FC6_Reboot, "harddrive": method.F14_Method, "ignoredisk": ignoredisk.F14_IgnoreDisk, "install": upgrade.F11_Upgrade, "iscsi": iscsi.F10_Iscsi, "iscsiname": iscsiname.FC6_IscsiName, "keyboard": keyboard.FC3_Keyboard, "lang": lang.FC3_Lang, "logging": logging.FC6_Logging, "logvol": logvol.F15_LogVol, "mediacheck": mediacheck.FC4_MediaCheck, "monitor": monitor.F10_Monitor, "multipath": multipath.FC6_MultiPath, "network": network.F9_Network, "nfs": method.F14_Method, "part": partition.F14_Partition, "partition": partition.F14_Partition, "poweroff": reboot.FC6_Reboot, "raid": raid.F15_Raid, "reboot": reboot.FC6_Reboot, "repo": repo.F15_Repo, "rescue": rescue.F10_Rescue, "rootpw": rootpw.F8_RootPw, "selinux": selinux.FC3_SELinux, "services": services.FC6_Services, "shutdown": reboot.FC6_Reboot, "skipx": skipx.FC3_SkipX, "sshpw": sshpw.F13_SshPw, "text": displaymode.FC3_DisplayMode, "timezone": timezone.FC6_Timezone, "updates": updates.F7_Updates, "upgrade": upgrade.F11_Upgrade, "url": method.F14_Method, "user": user.F12_User, "vnc": vnc.F9_Vnc, "volgroup": volgroup.FC3_VolGroup, "xconfig": xconfig.F14_XConfig, "zerombr": zerombr.F9_ZeroMbr, "zfcp": zfcp.F14_ZFCP, }, # based on f15 F16: { "auth": authconfig.FC3_Authconfig, "authconfig": authconfig.FC3_Authconfig, "autopart": autopart.F12_AutoPart, "autostep": autostep.FC3_AutoStep, "bootloader": bootloader.F15_Bootloader, "cdrom": method.F14_Method, "clearpart": clearpart.FC3_ClearPart, "cmdline": displaymode.FC3_DisplayMode, "device": device.F8_Device, "deviceprobe": deviceprobe.FC3_DeviceProbe, "dmraid": dmraid.FC6_DmRaid, "driverdisk": driverdisk.F14_DriverDisk, "fcoe": fcoe.F13_Fcoe, "firewall": firewall.F14_Firewall, "firstboot": firstboot.FC3_Firstboot, "graphical": displaymode.FC3_DisplayMode, "group": group.F12_Group, "halt": reboot.FC6_Reboot, "harddrive": method.F14_Method, "ignoredisk": ignoredisk.F14_IgnoreDisk, "install": upgrade.F11_Upgrade, "iscsi": iscsi.F10_Iscsi, "iscsiname": iscsiname.FC6_IscsiName, "keyboard": keyboard.FC3_Keyboard, "lang": lang.FC3_Lang, "logging": logging.FC6_Logging, "logvol": logvol.F15_LogVol, "mediacheck": mediacheck.FC4_MediaCheck, "monitor": monitor.F10_Monitor, "multipath": multipath.FC6_MultiPath, "network": network.F16_Network, "nfs": method.F14_Method, "part": partition.F14_Partition, "partition": partition.F14_Partition, "poweroff": reboot.FC6_Reboot, "raid": raid.F15_Raid, "reboot": reboot.FC6_Reboot, "repo": repo.F15_Repo, "rescue": rescue.F10_Rescue, "rootpw": rootpw.F8_RootPw, "selinux": selinux.FC3_SELinux, "services": services.FC6_Services, "shutdown": reboot.FC6_Reboot, "skipx": skipx.FC3_SkipX, "sshpw": sshpw.F13_SshPw, "text": displaymode.FC3_DisplayMode, "timezone": timezone.FC6_Timezone, "updates": updates.F7_Updates, "upgrade": upgrade.F11_Upgrade, "url": method.F14_Method, "user": user.F12_User, "vnc": vnc.F9_Vnc, "volgroup": volgroup.FC3_VolGroup, "xconfig": xconfig.F14_XConfig, "zerombr": zerombr.F9_ZeroMbr, "zfcp": zfcp.F14_ZFCP, }, # based on fc1 RHEL3: { "auth": authconfig.FC3_Authconfig, "authconfig": authconfig.FC3_Authconfig, "autopart": autopart.FC3_AutoPart, "autostep": autostep.FC3_AutoStep, "bootloader": bootloader.FC3_Bootloader, "cdrom": method.FC3_Method, "clearpart": clearpart.FC3_ClearPart, "cmdline": displaymode.FC3_DisplayMode, "device": device.FC3_Device, "deviceprobe": deviceprobe.FC3_DeviceProbe, "driverdisk": driverdisk.FC3_DriverDisk, "firewall": firewall.FC3_Firewall, "firstboot": firstboot.FC3_Firstboot, "graphical": displaymode.FC3_DisplayMode, "halt": reboot.FC3_Reboot, "harddrive": method.FC3_Method, "ignoredisk": ignoredisk.FC3_IgnoreDisk, "install": upgrade.FC3_Upgrade, "interactive": interactive.FC3_Interactive, "keyboard": keyboard.FC3_Keyboard, "lang": lang.FC3_Lang, "langsupport": langsupport.FC3_LangSupport, "lilo": bootloader.FC3_Bootloader, "lilocheck": lilocheck.FC3_LiloCheck, "logvol": logvol.FC3_LogVol, "monitor": monitor.FC3_Monitor, "mouse": mouse.RHEL3_Mouse, "network": network.FC3_Network, "nfs": method.FC3_Method, "part": partition.FC3_Partition, "partition": partition.FC3_Partition, "poweroff": reboot.FC3_Reboot, "raid": raid.FC3_Raid, "reboot": reboot.FC3_Reboot, "rootpw": rootpw.FC3_RootPw, "shutdown": reboot.FC3_Reboot, "skipx": skipx.FC3_SkipX, "text": displaymode.FC3_DisplayMode, "timezone": timezone.FC3_Timezone, "upgrade": upgrade.FC3_Upgrade, "url": method.FC3_Method, "vnc": vnc.FC3_Vnc, "volgroup": volgroup.FC3_VolGroup, "xconfig": xconfig.FC3_XConfig, "zerombr": zerombr.FC3_ZeroMbr, }, # based on fc3 RHEL4: { "auth": authconfig.FC3_Authconfig, "authconfig": authconfig.FC3_Authconfig, "autopart": autopart.FC3_AutoPart, "autostep": autostep.FC3_AutoStep, "bootloader": bootloader.FC3_Bootloader, "cdrom": method.FC3_Method, "clearpart": clearpart.FC3_ClearPart, "cmdline": displaymode.FC3_DisplayMode, "device": device.FC3_Device, "deviceprobe": deviceprobe.FC3_DeviceProbe, "driverdisk": driverdisk.FC4_DriverDisk, "firewall": firewall.FC3_Firewall, "firstboot": firstboot.FC3_Firstboot, "graphical": displaymode.FC3_DisplayMode, "halt": reboot.FC3_Reboot, "harddrive": method.FC3_Method, "ignoredisk": ignoredisk.F8_IgnoreDisk, "install": upgrade.FC3_Upgrade, "interactive": interactive.FC3_Interactive, "keyboard": keyboard.FC3_Keyboard, "lang": lang.FC3_Lang, "langsupport": langsupport.FC3_LangSupport, "lilo": bootloader.FC3_Bootloader, "lilocheck": lilocheck.FC3_LiloCheck, "logvol": logvol.FC3_LogVol, "monitor": monitor.FC3_Monitor, "mouse": mouse.FC3_Mouse, "network": network.RHEL4_Network, "nfs": method.FC3_Method, "part": partition.FC3_Partition, "partition": partition.FC3_Partition, "poweroff": reboot.FC3_Reboot, "raid": raid.FC3_Raid, "reboot": reboot.FC3_Reboot, "rootpw": rootpw.FC3_RootPw, "selinux": selinux.FC3_SELinux, "shutdown": reboot.FC3_Reboot, "skipx": skipx.FC3_SkipX, "text": displaymode.FC3_DisplayMode, "timezone": timezone.FC3_Timezone, "upgrade": upgrade.FC3_Upgrade, "url": method.FC3_Method, "vnc": vnc.FC3_Vnc, "volgroup": volgroup.FC3_VolGroup, "xconfig": xconfig.FC3_XConfig, "zerombr": zerombr.FC3_ZeroMbr, "zfcp": zfcp.FC3_ZFCP, }, # based on fc6 RHEL5: { "auth": authconfig.FC3_Authconfig, "authconfig": authconfig.FC3_Authconfig, "autopart": autopart.F9_AutoPart, "autostep": autostep.FC3_AutoStep, "bootloader": bootloader.RHEL5_Bootloader, "cdrom": method.FC6_Method, "clearpart": clearpart.FC3_ClearPart, "cmdline": displaymode.FC3_DisplayMode, "device": device.FC3_Device, "deviceprobe": deviceprobe.FC3_DeviceProbe, "dmraid": dmraid.FC6_DmRaid, "driverdisk": driverdisk.F12_DriverDisk, "firewall": firewall.FC3_Firewall, "firstboot": firstboot.FC3_Firstboot, "graphical": displaymode.FC3_DisplayMode, "halt": reboot.FC6_Reboot, "harddrive": method.FC6_Method, "ignoredisk": ignoredisk.F8_IgnoreDisk, "install": upgrade.FC3_Upgrade, "interactive": interactive.FC3_Interactive, "iscsi": iscsi.FC6_Iscsi, "iscsiname": iscsiname.FC6_IscsiName, "key": key.RHEL5_Key, "keyboard": keyboard.FC3_Keyboard, "lang": lang.FC3_Lang, "langsupport": langsupport.FC5_LangSupport, "logging": logging.FC6_Logging, "logvol": logvol.RHEL5_LogVol, "mediacheck": mediacheck.FC4_MediaCheck, "monitor": monitor.FC6_Monitor, "mouse": mouse.FC3_Mouse, "multipath": multipath.FC6_MultiPath, "network": network.RHEL5_Network, "nfs": method.FC6_Method, "part": partition.RHEL5_Partition, "partition": partition.RHEL5_Partition, "poweroff": reboot.FC6_Reboot, "raid": raid.RHEL5_Raid, "reboot": reboot.FC6_Reboot, "repo": repo.FC6_Repo, "rootpw": rootpw.FC3_RootPw, "services": services.FC6_Services, "selinux": selinux.FC3_SELinux, "shutdown": reboot.FC6_Reboot, "skipx": skipx.FC3_SkipX, "text": displaymode.FC3_DisplayMode, "timezone": timezone.FC6_Timezone, "upgrade": upgrade.FC3_Upgrade, "user": user.FC6_User, "url": method.FC6_Method, "vnc": vnc.FC6_Vnc, "volgroup": volgroup.FC3_VolGroup, "xconfig": xconfig.FC6_XConfig, "zerombr": zerombr.FC3_ZeroMbr, "zfcp": zfcp.FC3_ZFCP, }, # based on f13ish RHEL6: { "auth": authconfig.FC3_Authconfig, "authconfig": authconfig.FC3_Authconfig, "autopart": autopart.F12_AutoPart, "autostep": autostep.FC3_AutoStep, "bootloader": bootloader.RHEL6_Bootloader, "cdrom": method.RHEL6_Method, "clearpart": clearpart.FC3_ClearPart, "cmdline": displaymode.FC3_DisplayMode, "device": device.F8_Device, "deviceprobe": deviceprobe.FC3_DeviceProbe, "dmraid": dmraid.FC6_DmRaid, "driverdisk": driverdisk.F12_DriverDisk, "fcoe": fcoe.F13_Fcoe, "firewall": firewall.F10_Firewall, "firstboot": firstboot.FC3_Firstboot, "graphical": displaymode.FC3_DisplayMode, "group": group.F12_Group, "halt": reboot.FC6_Reboot, "harddrive": method.RHEL6_Method, "ignoredisk": ignoredisk.RHEL6_IgnoreDisk, "install": upgrade.F11_Upgrade, "interactive": interactive.FC3_Interactive, "iscsi": iscsi.F10_Iscsi, "iscsiname": iscsiname.FC6_IscsiName, "keyboard": keyboard.FC3_Keyboard, "lang": lang.FC3_Lang, "logging": logging.FC6_Logging, "logvol": logvol.F12_LogVol, "mediacheck": mediacheck.FC4_MediaCheck, "monitor": monitor.F10_Monitor, "multipath": multipath.FC6_MultiPath, "network": network.RHEL6_Network, "nfs": method.RHEL6_Method, "part": partition.F12_Partition, "partition": partition.F12_Partition, "poweroff": reboot.FC6_Reboot, "raid": raid.F13_Raid, "reboot": reboot.FC6_Reboot, "repo": repo.RHEL6_Repo, "rescue": rescue.F10_Rescue, "rootpw": rootpw.F8_RootPw, "selinux": selinux.FC3_SELinux, "services": services.FC6_Services, "shutdown": reboot.FC6_Reboot, "skipx": skipx.FC3_SkipX, "sshpw": sshpw.F13_SshPw, "text": displaymode.FC3_DisplayMode, "timezone": timezone.FC6_Timezone, "updates": updates.F7_Updates, "upgrade": upgrade.F11_Upgrade, "url": method.RHEL6_Method, "user": user.F12_User, "vnc": vnc.F9_Vnc, "volgroup": volgroup.FC3_VolGroup, "xconfig": xconfig.F10_XConfig, "zerombr": zerombr.F9_ZeroMbr, "zfcp": zfcp.F12_ZFCP, } } # This map is keyed on kickstart syntax version as provided by # pykickstart.version. Within each sub-dict is a mapping from a data object # name to the class that provides it. This is a bijective mapping - that is, # each name maps to exactly one data class and all data classes have a name. # More than one instance of each class is allowed to exist, however. dataMap = { FC3: { "DriverDiskData": driverdisk.FC3_DriverDiskData, "LogVolData": logvol.FC3_LogVolData, "NetworkData": network.FC3_NetworkData, "PartData": partition.FC3_PartData, "RaidData": raid.FC3_RaidData, "VolGroupData": volgroup.FC3_VolGroupData, "ZFCPData": zfcp.FC3_ZFCPData, }, FC4: { "DriverDiskData": driverdisk.FC4_DriverDiskData, "LogVolData": logvol.FC4_LogVolData, "NetworkData": network.FC4_NetworkData, "PartData": partition.FC4_PartData, "RaidData": raid.FC4_RaidData, "VolGroupData": volgroup.FC3_VolGroupData, "ZFCPData": zfcp.FC3_ZFCPData, }, FC5: { "DriverDiskData": driverdisk.FC4_DriverDiskData, "LogVolData": logvol.FC4_LogVolData, "NetworkData": network.FC4_NetworkData, "PartData": partition.FC4_PartData, "RaidData": raid.FC5_RaidData, "VolGroupData": volgroup.FC3_VolGroupData, "ZFCPData": zfcp.FC3_ZFCPData, }, FC6: { "DriverDiskData": driverdisk.FC4_DriverDiskData, "DmRaidData": dmraid.FC6_DmRaidData, "IscsiData": iscsi.FC6_IscsiData, "LogVolData": logvol.FC4_LogVolData, "MultiPathData": multipath.FC6_MultiPathData, "NetworkData": network.FC6_NetworkData, "PartData": partition.FC4_PartData, "RaidData": raid.FC5_RaidData, "RepoData": repo.FC6_RepoData, "UserData": user.FC6_UserData, "VolGroupData": volgroup.FC3_VolGroupData, "ZFCPData": zfcp.FC3_ZFCPData, }, F7: { "DriverDiskData": driverdisk.FC4_DriverDiskData, "DmRaidData": dmraid.FC6_DmRaidData, "IscsiData": iscsi.FC6_IscsiData, "LogVolData": logvol.FC4_LogVolData, "MultiPathData": multipath.FC6_MultiPathData, "NetworkData": network.FC6_NetworkData, "PartData": partition.FC4_PartData, "RaidData": raid.F7_RaidData, "RepoData": repo.FC6_RepoData, "UserData": user.FC6_UserData, "VolGroupData": volgroup.FC3_VolGroupData, "ZFCPData": zfcp.FC3_ZFCPData, }, F8: { "DriverDiskData": driverdisk.FC4_DriverDiskData, "DeviceData": device.F8_DeviceData, "DmRaidData": dmraid.FC6_DmRaidData, "IscsiData": iscsi.FC6_IscsiData, "LogVolData": logvol.FC4_LogVolData, "MultiPathData": multipath.FC6_MultiPathData, "NetworkData": network.F8_NetworkData, "PartData": partition.FC4_PartData, "RaidData": raid.F7_RaidData, "RepoData": repo.F8_RepoData, "UserData": user.F8_UserData, "VolGroupData": volgroup.FC3_VolGroupData, "ZFCPData": zfcp.FC3_ZFCPData, }, F9: { "DriverDiskData": driverdisk.FC4_DriverDiskData, "DeviceData": device.F8_DeviceData, "DmRaidData": dmraid.FC6_DmRaidData, "IscsiData": iscsi.FC6_IscsiData, "LogVolData": logvol.F9_LogVolData, "MultiPathData": multipath.FC6_MultiPathData, "NetworkData": network.F8_NetworkData, "PartData": partition.F9_PartData, "RaidData": raid.F9_RaidData, "RepoData": repo.F8_RepoData, "UserData": user.F8_UserData, "VolGroupData": volgroup.FC3_VolGroupData, "ZFCPData": zfcp.FC3_ZFCPData, }, F10: { "DriverDiskData": driverdisk.FC4_DriverDiskData, "DeviceData": device.F8_DeviceData, "DmRaidData": dmraid.FC6_DmRaidData, "IscsiData": iscsi.F10_IscsiData, "LogVolData": logvol.F9_LogVolData, "MultiPathData": multipath.FC6_MultiPathData, "NetworkData": network.F8_NetworkData, "PartData": partition.F9_PartData, "RaidData": raid.F9_RaidData, "RepoData": repo.F8_RepoData, "UserData": user.F8_UserData, "VolGroupData": volgroup.FC3_VolGroupData, "ZFCPData": zfcp.FC3_ZFCPData, }, F11: { "DriverDiskData": driverdisk.FC4_DriverDiskData, "DeviceData": device.F8_DeviceData, "DmRaidData": dmraid.FC6_DmRaidData, "IscsiData": iscsi.F10_IscsiData, "LogVolData": logvol.F9_LogVolData, "MultiPathData": multipath.FC6_MultiPathData, "NetworkData": network.F8_NetworkData, "PartData": partition.F11_PartData, "RaidData": raid.F9_RaidData, "RepoData": repo.F11_RepoData, "UserData": user.F8_UserData, "VolGroupData": volgroup.FC3_VolGroupData, "ZFCPData": zfcp.FC3_ZFCPData, }, F12: { "DriverDiskData": driverdisk.F12_DriverDiskData, "DeviceData": device.F8_DeviceData, "DmRaidData": dmraid.FC6_DmRaidData, "FcoeData": fcoe.F12_FcoeData, "GroupData": group.F12_GroupData, "IscsiData": iscsi.F10_IscsiData, "LogVolData": logvol.F12_LogVolData, "MultiPathData": multipath.FC6_MultiPathData, "NetworkData": network.F8_NetworkData, "PartData": partition.F12_PartData, "RaidData": raid.F12_RaidData, "RepoData": repo.F11_RepoData, "UserData": user.F12_UserData, "VolGroupData": volgroup.FC3_VolGroupData, "ZFCPData": zfcp.F12_ZFCPData, }, F13: { "DriverDiskData": driverdisk.F12_DriverDiskData, "DeviceData": device.F8_DeviceData, "DmRaidData": dmraid.FC6_DmRaidData, "FcoeData": fcoe.F13_FcoeData, "GroupData": group.F12_GroupData, "IscsiData": iscsi.F10_IscsiData, "LogVolData": logvol.F12_LogVolData, "MultiPathData": multipath.FC6_MultiPathData, "NetworkData": network.F8_NetworkData, "PartData": partition.F12_PartData, "RaidData": raid.F13_RaidData, "RepoData": repo.F13_RepoData, "SshPwData": sshpw.F13_SshPwData, "UserData": user.F12_UserData, "VolGroupData": volgroup.FC3_VolGroupData, "ZFCPData": zfcp.F12_ZFCPData, }, F14: { "DriverDiskData": driverdisk.F14_DriverDiskData, "DeviceData": device.F8_DeviceData, "DmRaidData": dmraid.FC6_DmRaidData, "FcoeData": fcoe.F13_FcoeData, "GroupData": group.F12_GroupData, "IscsiData": iscsi.F10_IscsiData, "LogVolData": logvol.F14_LogVolData, "MultiPathData": multipath.FC6_MultiPathData, "NetworkData": network.F8_NetworkData, "PartData": partition.F14_PartData, "RaidData": raid.F14_RaidData, "RepoData": repo.F14_RepoData, "SshPwData": sshpw.F13_SshPwData, "UserData": user.F12_UserData, "VolGroupData": volgroup.FC3_VolGroupData, "ZFCPData": zfcp.F14_ZFCPData, }, F15: { "DriverDiskData": driverdisk.F14_DriverDiskData, "DeviceData": device.F8_DeviceData, "DmRaidData": dmraid.FC6_DmRaidData, "FcoeData": fcoe.F13_FcoeData, "GroupData": group.F12_GroupData, "IscsiData": iscsi.F10_IscsiData, "LogVolData": logvol.F15_LogVolData, "MultiPathData": multipath.FC6_MultiPathData, "NetworkData": network.F8_NetworkData, "PartData": partition.F14_PartData, "RaidData": raid.F15_RaidData, "RepoData": repo.F15_RepoData, "SshPwData": sshpw.F13_SshPwData, "UserData": user.F12_UserData, "VolGroupData": volgroup.FC3_VolGroupData, "ZFCPData": zfcp.F14_ZFCPData, }, F16: { "DriverDiskData": driverdisk.F14_DriverDiskData, "DeviceData": device.F8_DeviceData, "DmRaidData": dmraid.FC6_DmRaidData, "FcoeData": fcoe.F13_FcoeData, "GroupData": group.F12_GroupData, "IscsiData": iscsi.F10_IscsiData, "LogVolData": logvol.F15_LogVolData, "MultiPathData": multipath.FC6_MultiPathData, "NetworkData": network.F16_NetworkData, "PartData": partition.F14_PartData, "RaidData": raid.F15_RaidData, "RepoData": repo.F15_RepoData, "SshPwData": sshpw.F13_SshPwData, "UserData": user.F12_UserData, "VolGroupData": volgroup.FC3_VolGroupData, "ZFCPData": zfcp.F14_ZFCPData, }, RHEL3: { "DriverDiskData": driverdisk.FC3_DriverDiskData, "LogVolData": logvol.FC3_LogVolData, "NetworkData": network.RHEL4_NetworkData, "PartData": partition.FC3_PartData, "RaidData": raid.FC3_RaidData, "VolGroupData": volgroup.FC3_VolGroupData, "ZFCPData": zfcp.FC3_ZFCPData, }, RHEL4: { "DriverDiskData": driverdisk.FC4_DriverDiskData, "LogVolData": logvol.FC3_LogVolData, "NetworkData": network.RHEL4_NetworkData, "PartData": partition.FC3_PartData, "RaidData": raid.FC3_RaidData, "VolGroupData": volgroup.FC3_VolGroupData, "ZFCPData": zfcp.FC3_ZFCPData, }, RHEL5: { "DriverDiskData": driverdisk.F12_DriverDiskData, "DmRaidData": dmraid.FC6_DmRaidData, "IscsiData": iscsi.FC6_IscsiData, "LogVolData": logvol.RHEL5_LogVolData, "MultiPathData": multipath.FC6_MultiPathData, "NetworkData": network.FC6_NetworkData, "PartData": partition.RHEL5_PartData, "RaidData": raid.RHEL5_RaidData, "RepoData": repo.FC6_RepoData, "UserData": user.FC6_UserData, "VolGroupData": volgroup.FC3_VolGroupData, "ZFCPData": zfcp.FC3_ZFCPData, }, RHEL6: { "DriverDiskData": driverdisk.F12_DriverDiskData, "DeviceData": device.F8_DeviceData, "DmRaidData": dmraid.FC6_DmRaidData, "FcoeData": fcoe.F13_FcoeData, "GroupData": group.F12_GroupData, "IscsiData": iscsi.F10_IscsiData, "LogVolData": logvol.F12_LogVolData, "MultiPathData": multipath.FC6_MultiPathData, "NetworkData": network.RHEL6_NetworkData, "PartData": partition.F12_PartData, "RaidData": raid.F13_RaidData, "RepoData": repo.RHEL6_RepoData, "SshPwData": sshpw.F13_SshPwData, "UserData": user.F12_UserData, "VolGroupData": volgroup.FC3_VolGroupData, "ZFCPData": zfcp.F12_ZFCPData, } } ########NEW FILE######## __FILENAME__ = f10 # # Chris Lumens <clumens@redhat.com> # # Copyright 2008 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.version import * class F10Handler(BaseHandler): version = F10 ########NEW FILE######## __FILENAME__ = f11 # # Chris Lumens <clumens@redhat.com> # # Copyright 2008 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.version import * class F11Handler(BaseHandler): version = F11 ########NEW FILE######## __FILENAME__ = f12 # # Chris Lumens <clumens@redhat.com> # # Copyright 2009 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.version import * class F12Handler(BaseHandler): version = F12 ########NEW FILE######## __FILENAME__ = f13 # # Chris Lumens <clumens@redhat.com> # # Copyright 2009 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.version import * class F13Handler(BaseHandler): version = F13 ########NEW FILE######## __FILENAME__ = f14 # # Chris Lumens <clumens@redhat.com> # # Copyright 2010 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.version import * class F14Handler(BaseHandler): version = F14 ########NEW FILE######## __FILENAME__ = f15 # # Chris Lumens <clumens@redhat.com> # # Copyright 2010 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.version import * class F15Handler(BaseHandler): version = F15 ########NEW FILE######## __FILENAME__ = f16 # # Chris Lumens <clumens@redhat.com> # # Copyright 2011 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.version import * class F16Handler(BaseHandler): version = F16 ########NEW FILE######## __FILENAME__ = f7 # # Chris Lumens <clumens@redhat.com> # # Copyright 2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.version import * class F7Handler(BaseHandler): version = F7 ########NEW FILE######## __FILENAME__ = f8 # # Chris Lumens <clumens@redhat.com> # # Copyright 2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.version import * class F8Handler(BaseHandler): version = F8 ########NEW FILE######## __FILENAME__ = f9 # # Chris Lumens <clumens@redhat.com> # # Copyright 2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.version import * class F9Handler(BaseHandler): version = F9 ########NEW FILE######## __FILENAME__ = fc3 # # Chris Lumens <clumens@redhat.com> # # Copyright 2005, 2006, 2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.version import * class FC3Handler(BaseHandler): version = FC3 ########NEW FILE######## __FILENAME__ = fc4 # # Chris Lumens <clumens@redhat.com> # # Copyright 2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.version import * class FC4Handler(BaseHandler): version = FC4 ########NEW FILE######## __FILENAME__ = fc5 # # Chris Lumens <clumens@redhat.com> # # Copyright 2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.version import * class FC5Handler(BaseHandler): version = FC5 ########NEW FILE######## __FILENAME__ = fc6 # # Chris Lumens <clumens@redhat.com> # # Copyright 2006, 2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.version import * class FC6Handler(BaseHandler): version = FC6 ########NEW FILE######## __FILENAME__ = rhel3 # # Chris Lumens <clumens@redhat.com> # # Copyright 2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.version import * class RHEL3Handler(BaseHandler): version = RHEL3 ########NEW FILE######## __FILENAME__ = rhel4 # # Chris Lumens <clumens@redhat.com> # # Copyright 2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.version import * class RHEL4Handler(BaseHandler): version = RHEL4 ########NEW FILE######## __FILENAME__ = rhel5 # # Chris Lumens <clumens@redhat.com> # # Copyright 2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.version import * class RHEL5Handler(BaseHandler): version = RHEL5 ########NEW FILE######## __FILENAME__ = rhel6 # # Chris Lumens <clumens@redhat.com> # # Copyright 2010 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # from pykickstart.base import * from pykickstart.version import * class RHEL6Handler(BaseHandler): version = RHEL6 ########NEW FILE######## __FILENAME__ = ko # # Chris Lumens <clumens@redhat.com> # # Copyright 2009 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # """ Base classes for internal pykickstart use. The module exports the following important classes: KickstartObject - The base class for all classes in pykickstart """ class KickstartObject(object): """The base class for all other classes in pykickstart.""" def __init__(self, *args, **kwargs): """Create a new KickstartObject instance. All other classes in pykickstart should be derived from this one. Instance attributes: """ pass def __str__(self): return "" ########NEW FILE######## __FILENAME__ = options # # Chris Lumens <clumens@redhat.com> # # Copyright 2005, 2006, 2007 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # """ Specialized option handling. This module exports two classes: KSOptionParser - A specialized subclass of OptionParser to be used in BaseHandler subclasses. KSOption - A specialized subclass of Option. """ import warnings from copy import copy from optparse import * from constants import * from errors import * from version import * import gettext _ = lambda x: gettext.ldgettext("pykickstart", x) class KSOptionParser(OptionParser): """A specialized subclass of optparse.OptionParser to handle extra option attribute checking, work error reporting into the KickstartParseError framework, and to turn off the default help. """ def exit(self, status=0, msg=None): pass def error(self, msg): if self.lineno != None: raise KickstartParseError, formatErrorMsg(self.lineno, msg=msg) else: raise KickstartParseError, msg def keys(self): retval = [] for opt in self.option_list: if opt not in retval: retval.append(opt.dest) return retval def _init_parsing_state (self): OptionParser._init_parsing_state(self) self.option_seen = {} def check_values (self, values, args): def seen(self, option): return self.option_seen.has_key(option) def usedTooNew(self, option): return option.introduced and option.introduced > self.version def usedDeprecated(self, option): return option.deprecated def usedRemoved(self, option): return option.removed and option.removed <= self.version for option in filter(lambda o: isinstance(o, Option), self.option_list): if option.required and not seen(self, option): raise KickstartValueError, formatErrorMsg(self.lineno, _("Option %s is required") % option) elif seen(self, option) and usedTooNew(self, option): mapping = {"option": option, "intro": versionToString(option.introduced), "version": versionToString(self.version)} self.error(_("The %(option)s option was introduced in version %(intro)s, but you are using kickstart syntax version %(version)s.") % mapping) elif seen(self, option) and usedRemoved(self, option): mapping = {"option": option, "removed": versionToString(option.removed), "version": versionToString(self.version)} if option.removed == self.version: self.error(_("The %(option)s option is no longer supported.") % mapping) else: self.error(_("The %(option)s option was removed in version %(removed)s, but you are using kickstart syntax version %(version)s.") % mapping) elif seen(self, option) and usedDeprecated(self, option): mapping = {"lineno": self.lineno, "option": option} warnings.warn(_("Ignoring deprecated option on line %(lineno)s: The %(option)s option has been deprecated and no longer has any effect. It may be removed from future releases, which will result in a fatal error from kickstart. Please modify your kickstart file to remove this option.") % mapping, DeprecationWarning) return (values, args) def parse_args(self, *args, **kwargs): if kwargs.has_key("lineno"): self.lineno = kwargs.pop("lineno") return OptionParser.parse_args(self, **kwargs) def __init__(self, mapping=None, version=None): """Create a new KSOptionParser instance. Each KickstartCommand subclass should create one instance of KSOptionParser, providing at least the lineno attribute. mapping and version are not required. Instance attributes: mapping -- A mapping from option strings to different values. version -- The version of the kickstart syntax we are checking against. """ OptionParser.__init__(self, option_class=KSOption, add_help_option=False, conflict_handler="resolve") if mapping is None: self.map = {} else: self.map = mapping self.lineno = None self.option_seen = {} self.version = version def _check_ksboolean(option, opt, value): if value.lower() in ("on", "yes", "true", "1"): return True elif value.lower() in ("off", "no", "false", "0"): return False else: mapping = {"opt": opt, "value": value} raise OptionValueError(_("Option %(opt)s: invalid boolean value: %(value)r") % mapping) def _check_string(option, opt, value): if len(value) > 2 and value.startswith("--"): mapping = {"opt": opt, "value": value} raise OptionValueError(_("Option %(opt)s: invalid string value: %(value)r") % mapping) else: return value # Creates a new Option class that supports several new attributes: # - required: any option with this attribute must be supplied or an exception # is thrown # - introduced: the kickstart syntax version that this option first appeared # in - an exception will be raised if the option is used and # the specified syntax version is less than the value of this # attribute # - deprecated: the kickstart syntax version that this option was deprecated # in - a DeprecationWarning will be thrown if the option is # used and the specified syntax version is greater than the # value of this attribute # - removed: the kickstart syntax version that this option was removed in - an # exception will be raised if the option is used and the specified # syntax version is greated than the value of this attribute # Also creates a new type: # - ksboolean: support various kinds of boolean values on an option # And two new actions: # - map : allows you to define an opt -> val mapping such that dest gets val # when opt is seen # - map_extend: allows you to define an opt -> [val1, ... valn] mapping such # that dest gets a list of vals built up when opt is seen class KSOption (Option): ATTRS = Option.ATTRS + ['introduced', 'deprecated', 'removed', 'required'] ACTIONS = Option.ACTIONS + ("map", "map_extend",) STORE_ACTIONS = Option.STORE_ACTIONS + ("map", "map_extend",) TYPES = Option.TYPES + ("ksboolean", "string") TYPE_CHECKER = copy(Option.TYPE_CHECKER) TYPE_CHECKER["ksboolean"] = _check_ksboolean TYPE_CHECKER["string"] = _check_string def _check_required(self): if self.required and not self.takes_value(): raise OptionError(_("Required flag set for option that doesn't take a value"), self) # Make sure _check_required() is called from the constructor! CHECK_METHODS = Option.CHECK_METHODS + [_check_required] def process (self, opt, value, values, parser): Option.process(self, opt, value, values, parser) parser.option_seen[self] = 1 # Override default take_action method to handle our custom actions. def take_action(self, action, dest, opt, value, values, parser): if action == "map": values.ensure_value(dest, parser.map[opt.lstrip('-')]) elif action == "map_extend": values.ensure_value(dest, []).extend(parser.map[opt.lstrip('-')]) else: Option.take_action(self, action, dest, opt, value, values, parser) def takes_value(self): # Deprecated options don't take a value. return Option.takes_value(self) and not self.deprecated def __init__(self, *args, **kwargs): self.deprecated = False self.required = False Option.__init__(self, *args, **kwargs) ########NEW FILE######## __FILENAME__ = parser # # parser.py: Kickstart file parser. # # Chris Lumens <clumens@redhat.com> # # Copyright 2005, 2006, 2007, 2008, 2011 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # """ Main kickstart file processing module. This module exports several important classes: Script - Representation of a single %pre, %post, or %traceback script. Packages - Representation of the %packages section. KickstartParser - The kickstart file parser state machine. """ from collections import Iterator import os import shlex import sys import tempfile from copy import copy from optparse import * from urlgrabber import urlread import urlgrabber.grabber as grabber import constants from errors import KickstartError, KickstartParseError, KickstartValueError, formatErrorMsg from ko import KickstartObject from sections import * import version import gettext _ = lambda x: gettext.ldgettext("pykickstart", x) STATE_END = "end" STATE_COMMANDS = "commands" ver = version.DEVEL def _preprocessStateMachine (lineIter): l = None lineno = 0 # Now open an output kickstart file that we are going to write to one # line at a time. (outF, outName) = tempfile.mkstemp("-ks.cfg", "", "/tmp") while True: try: l = lineIter.next() except StopIteration: break # At the end of the file? if l == "": break lineno += 1 url = None ll = l.strip() if not ll.startswith("%ksappend"): os.write(outF, l) continue # Try to pull down the remote file. try: ksurl = ll.split(' ')[1] except: raise KickstartParseError, formatErrorMsg(lineno, msg=_("Illegal url for %%ksappend: %s") % ll) try: url = grabber.urlopen(ksurl) except grabber.URLGrabError, e: raise KickstartError, formatErrorMsg(lineno, msg=_("Unable to open %%ksappend file: %s") % e.strerror) else: # Sanity check result. Sometimes FTP doesn't catch a file # is missing. try: if url.size < 1: raise KickstartError, formatErrorMsg(lineno, msg=_("Unable to open %%ksappend file")) except: raise KickstartError, formatErrorMsg(lineno, msg=_("Unable to open %%ksappend file")) # If that worked, write the remote file to the output kickstart # file in one burst. Then close everything up to get ready to # read ahead in the input file. This allows multiple %ksappend # lines to exist. if url is not None: os.write(outF, url.read()) url.close() # All done - close the temp file and return its location. os.close(outF) return outName def preprocessFromString (s): """Preprocess the kickstart file, provided as the string str. This method is currently only useful for handling %ksappend lines, which need to be fetched before the real kickstart parser can be run. Returns the location of the complete kickstart file. """ i = iter(s.splitlines(True) + [""]) rc = _preprocessStateMachine (i.next) return rc def preprocessKickstart (f): """Preprocess the kickstart file, given by the filename file. This method is currently only useful for handling %ksappend lines, which need to be fetched before the real kickstart parser can be run. Returns the location of the complete kickstart file. """ try: fh = urlopen(f) except grabber.URLGrabError, e: raise KickstartError, formatErrorMsg(0, msg=_("Unable to open input kickstart file: %s") % e.strerror) rc = _preprocessStateMachine (iter(fh.readlines())) fh.close() return rc class PutBackIterator(Iterator): def __init__(self, iterable): self._iterable = iter(iterable) self._buf = None def __iter__(self): return self def put(self, s): self._buf = s def next(self): if self._buf: retval = self._buf self._buf = None return retval else: return self._iterable.next() ### ### SCRIPT HANDLING ### class Script(KickstartObject): """A class representing a single kickstart script. If functionality beyond just a data representation is needed (for example, a run method in anaconda), Script may be subclassed. Although a run method is not provided, most of the attributes of Script have to do with running the script. Instances of Script are held in a list by the Version object. """ def __init__(self, script, *args , **kwargs): """Create a new Script instance. Instance attributes: errorOnFail -- If execution of the script fails, should anaconda stop, display an error, and then reboot without running any other scripts? inChroot -- Does the script execute in anaconda's chroot environment or not? interp -- The program that should be used to interpret this script. lineno -- The line number this script starts on. logfile -- Where all messages from the script should be logged. script -- A string containing all the lines of the script. type -- The type of the script, which can be KS_SCRIPT_* from pykickstart.constants. """ KickstartObject.__init__(self, *args, **kwargs) self.script = "".join(script) self.interp = kwargs.get("interp", "/bin/sh") self.inChroot = kwargs.get("inChroot", False) self.lineno = kwargs.get("lineno", None) self.logfile = kwargs.get("logfile", None) self.errorOnFail = kwargs.get("errorOnFail", False) self.type = kwargs.get("type", constants.KS_SCRIPT_PRE) def __str__(self): """Return a string formatted for output to a kickstart file.""" retval = "" if self.type == constants.KS_SCRIPT_PRE: retval += '\n%pre' elif self.type == constants.KS_SCRIPT_POST: retval += '\n%post' elif self.type == constants.KS_SCRIPT_TRACEBACK: retval += '\n%traceback' if self.interp != "/bin/sh" and self.interp != "": retval += " --interpreter=%s" % self.interp if self.type == constants.KS_SCRIPT_POST and not self.inChroot: retval += " --nochroot" if self.logfile != None: retval += " --logfile %s" % self.logfile if self.errorOnFail: retval += " --erroronfail" if self.script.endswith("\n"): if ver >= version.F8: return retval + "\n%s%%end\n" % self.script else: return retval + "\n%s\n" % self.script else: if ver >= version.F8: return retval + "\n%s\n%%end\n" % self.script else: return retval + "\n%s\n" % self.script ## ## PACKAGE HANDLING ## class Group: """A class representing a single group in the %packages section.""" def __init__(self, name="", include=constants.GROUP_DEFAULT): """Create a new Group instance. Instance attributes: name -- The group's identifier include -- The level of how much of the group should be included. Values can be GROUP_* from pykickstart.constants. """ self.name = name self.include = include def __str__(self): """Return a string formatted for output to a kickstart file.""" if self.include == constants.GROUP_REQUIRED: return "@%s --nodefaults" % self.name elif self.include == constants.GROUP_ALL: return "@%s --optional" % self.name else: return "@%s" % self.name def __cmp__(self, other): if self.name < other.name: return -1 elif self.name > other.name: return 1 return 0 class Packages(KickstartObject): """A class representing the %packages section of the kickstart file.""" def __init__(self, *args, **kwargs): """Create a new Packages instance. Instance attributes: addBase -- Should the Base group be installed even if it is not specified? default -- Should the default package set be selected? excludedList -- A list of all the packages marked for exclusion in the %packages section, without the leading minus symbol. excludeDocs -- Should documentation in each package be excluded? groupList -- A list of Group objects representing all the groups specified in the %packages section. Names will be stripped of the leading @ symbol. excludedGroupList -- A list of Group objects representing all the groups specified for removal in the %packages section. Names will be stripped of the leading -@ symbols. handleMissing -- If unknown packages are specified in the %packages section, should it be ignored or not? Values can be KS_MISSING_* from pykickstart.constants. packageList -- A list of all the packages specified in the %packages section. instLangs -- A list of languages to install. """ KickstartObject.__init__(self, *args, **kwargs) self.addBase = True self.default = False self.excludedList = [] self.excludedGroupList = [] self.excludeDocs = False self.groupList = [] self.handleMissing = constants.KS_MISSING_PROMPT self.packageList = [] self.instLangs = None def __str__(self): """Return a string formatted for output to a kickstart file.""" pkgs = "" if not self.default: grps = self.groupList grps.sort() for grp in grps: pkgs += "%s\n" % grp.__str__() p = self.packageList p.sort() for pkg in p: pkgs += "%s\n" % pkg grps = self.excludedGroupList grps.sort() for grp in grps: pkgs += "-%s\n" % grp.__str__() p = self.excludedList p.sort() for pkg in p: pkgs += "-%s\n" % pkg if pkgs == "": return "" retval = "\n%packages" if self.default: retval += " --default" if self.excludeDocs: retval += " --excludedocs" if not self.addBase: retval += " --nobase" if self.handleMissing == constants.KS_MISSING_IGNORE: retval += " --ignoremissing" if self.instLangs: retval += " --instLangs=%s" % self.instLangs if ver >= version.F8: return retval + "\n" + pkgs + "\n%end\n" else: return retval + "\n" + pkgs + "\n" def _processGroup (self, line): op = OptionParser() op.add_option("--nodefaults", action="store_true", default=False) op.add_option("--optional", action="store_true", default=False) (opts, extra) = op.parse_args(args=line.split()) if opts.nodefaults and opts.optional: raise KickstartValueError, _("Group cannot specify both --nodefaults and --optional") # If the group name has spaces in it, we have to put it back together # now. grp = " ".join(extra) if opts.nodefaults: self.groupList.append(Group(name=grp, include=constants.GROUP_REQUIRED)) elif opts.optional: self.groupList.append(Group(name=grp, include=constants.GROUP_ALL)) else: self.groupList.append(Group(name=grp, include=constants.GROUP_DEFAULT)) def add (self, pkgList): """Given a list of lines from the input file, strip off any leading symbols and add the result to the appropriate list. """ existingExcludedSet = set(self.excludedList) existingPackageSet = set(self.packageList) newExcludedSet = set() newPackageSet = set() excludedGroupList = [] for pkg in pkgList: stripped = pkg.strip() if stripped[0] == "@": self._processGroup(stripped[1:]) elif stripped[0] == "-": if stripped[1] == "@": excludedGroupList.append(Group(name=stripped[2:])) else: newExcludedSet.add(stripped[1:]) else: newPackageSet.add(stripped) # Groups have to be excluded in two different ways (note: can't use # sets here because we have to store objects): excludedGroupNames = map(lambda g: g.name, excludedGroupList) # First, an excluded group may be cancelling out a previously given # one. This is often the case when using %include. So there we should # just remove the group from the list. self.groupList = filter(lambda g: g.name not in excludedGroupNames, self.groupList) # Second, the package list could have included globs which are not # processed by pykickstart. In that case we need to preserve a list of # excluded groups so whatever tool doing package/group installation can # take appropriate action. self.excludedGroupList.extend(excludedGroupList) existingPackageSet = (existingPackageSet - newExcludedSet) | newPackageSet existingExcludedSet = (existingExcludedSet - existingPackageSet) | newExcludedSet self.packageList = list(existingPackageSet) self.excludedList = list(existingExcludedSet) ### ### PARSER ### class KickstartParser: """The kickstart file parser class as represented by a basic state machine. To create a specialized parser, make a subclass and override any of the methods you care about. Methods that don't need to do anything may just pass. However, _stateMachine should never be overridden. """ def __init__ (self, handler, followIncludes=True, errorsAreFatal=True, missingIncludeIsFatal=True): """Create a new KickstartParser instance. Instance attributes: errorsAreFatal -- Should errors cause processing to halt, or just print a message to the screen? This is most useful for writing syntax checkers that may want to continue after an error is encountered. followIncludes -- If %include is seen, should the included file be checked as well or skipped? handler -- An instance of a BaseHandler subclass. If None, the input file will still be parsed but no data will be saved and no commands will be executed. missingIncludeIsFatal -- Should missing include files be fatal, even if errorsAreFatal is False? """ self.errorsAreFatal = errorsAreFatal self.followIncludes = followIncludes self.handler = handler self.currentdir = {} self.missingIncludeIsFatal = missingIncludeIsFatal self._state = STATE_COMMANDS self._includeDepth = 0 self._line = "" self.version = self.handler.version global ver ver = self.version self._sections = {} self.setupSections() def _reset(self): """Reset the internal variables of the state machine for a new kickstart file.""" self._state = STATE_COMMANDS self._includeDepth = 0 def getSection(self, s): """Return a reference to the requested section (s must start with '%'s), or raise KeyError if not found. """ return self._sections[s] def handleCommand (self, lineno, args): """Given the list of command and arguments, call the Version's dispatcher method to handle the command. Returns the command or data object returned by the dispatcher. This method may be overridden in a subclass if necessary. """ if self.handler: self.handler.currentCmd = args[0] self.handler.currentLine = self._line retval = self.handler.dispatcher(args, lineno) return retval def registerSection(self, obj): """Given an instance of a Section subclass, register the new section with the parser. Calling this method means the parser will recognize your new section and dispatch into the given object to handle it. """ if not obj.sectionOpen: raise TypeError, "no sectionOpen given for section %s" % obj if not obj.sectionOpen.startswith("%"): raise TypeError, "section %s tag does not start with a %%" % obj.sectionOpen self._sections[obj.sectionOpen] = obj def _finalize(self, obj): """Called at the close of a kickstart section to take any required actions. Internally, this is used to add scripts once we have the whole body read. """ obj.finalize() self._state = STATE_COMMANDS def _handleSpecialComments(self, line): """Kickstart recognizes a couple special comments.""" if self._state != STATE_COMMANDS: return # Save the platform for s-c-kickstart. if line[:10] == "#platform=": self.handler.platform = self._line[11:] def _readSection(self, lineIter, lineno): obj = self._sections[self._state] while True: try: line = lineIter.next() if line == "": # This section ends at the end of the file. if self.version >= version.F8: raise KickstartParseError, formatErrorMsg(lineno, msg=_("Section does not end with %%end.")) self._finalize(obj) except StopIteration: break lineno += 1 # Throw away blank lines and comments, unless the section wants all # lines. if self._isBlankOrComment(line) and not obj.allLines: continue if line.startswith("%"): args = shlex.split(line) if args and args[0] == "%end": # This is a properly terminated section. self._finalize(obj) break elif args and args[0] == "%ksappend": continue elif args and (self._validState(args[0]) or args[0] in ["%include", "%ksappend"]): # This is an unterminated section. if self.version >= version.F8: raise KickstartParseError, formatErrorMsg(lineno, msg=_("Section does not end with %%end.")) # Finish up. We do not process the header here because # kicking back out to STATE_COMMANDS will ensure that happens. lineIter.put(line) lineno -= 1 self._finalize(obj) break else: # This is just a line within a section. Pass it off to whatever # section handles it. obj.handleLine(line) return lineno def _validState(self, st): """Is the given section tag one that has been registered with the parser?""" return st in self._sections.keys() def _tryFunc(self, fn): """Call the provided function (which doesn't take any arguments) and do the appropriate error handling. If errorsAreFatal is False, this function will just print the exception and keep going. """ try: fn() except Exception, msg: if self.errorsAreFatal: raise else: print msg def _isBlankOrComment(self, line): return line.isspace() or line == "" or line.lstrip()[0] == '#' def _stateMachine(self, lineIter): # For error reporting. lineno = 0 while True: # Get the next line out of the file, quitting if this is the last line. try: self._line = lineIter.next() if self._line == "": break except StopIteration: break lineno += 1 # Eliminate blank lines, whitespace-only lines, and comments. if self._isBlankOrComment(self._line): self._handleSpecialComments(self._line) continue # Remove any end-of-line comments. sanitized = self._line.split("#")[0] # Then split the line. args = shlex.split(sanitized.rstrip()) if args[0] == "%include": # This case comes up primarily in ksvalidator. if not self.followIncludes: continue if len(args) == 1 or not args[1]: raise KickstartParseError, formatErrorMsg(lineno) self._includeDepth += 1 try: self.readKickstart(args[1], reset=False) except KickstartError: # Handle the include file being provided over the # network in a %pre script. This case comes up in the # early parsing in anaconda. if self.missingIncludeIsFatal: raise self._includeDepth -= 1 continue # Now on to the main event. if self._state == STATE_COMMANDS: if args[0] == "%ksappend": # This is handled by the preprocess* functions, so continue. continue elif args[0][0] == '%': # This is the beginning of a new section. Handle its header # here. newSection = args[0] if not self._validState(newSection): raise KickstartParseError, formatErrorMsg(lineno, msg=_("Unknown kickstart section: %s" % newSection)) self._state = newSection obj = self._sections[self._state] self._tryFunc(lambda: obj.handleHeader(lineno, args)) # This will handle all section processing, kicking us back # out to STATE_COMMANDS at the end with the current line # being the next section header, etc. lineno = self._readSection(lineIter, lineno) else: # This is a command in the command section. Dispatch to it. self._tryFunc(lambda: self.handleCommand(lineno, args)) elif self._state == STATE_END: break def readKickstartFromString (self, s, reset=True): """Process a kickstart file, provided as the string str.""" if reset: self._reset() # Add a "" to the end of the list so the string reader acts like the # file reader and we only get StopIteration when we're after the final # line of input. i = PutBackIterator(s.splitlines(True) + [""]) self._stateMachine (i) def readKickstart(self, f, reset=True): """Process a kickstart file, given by the filename f.""" if reset: self._reset() # an %include might not specify a full path. if we don't try to figure # out what the path should have been, then we're unable to find it # requiring full path specification, though, sucks. so let's make # the reading "smart" by keeping track of what the path is at each # include depth. if not os.path.exists(f): if self.currentdir.has_key(self._includeDepth - 1): if os.path.exists(os.path.join(self.currentdir[self._includeDepth - 1], f)): f = os.path.join(self.currentdir[self._includeDepth - 1], f) cd = os.path.dirname(f) if not cd.startswith("/"): cd = os.path.abspath(cd) self.currentdir[self._includeDepth] = cd try: s = urlread(f) except grabber.URLGrabError, e: raise KickstartError, formatErrorMsg(0, msg=_("Unable to open input kickstart file: %s") % e.strerror) self.readKickstartFromString(s, reset=False) def setupSections(self): """Install the sections all kickstart files support. You may override this method in a subclass, but should avoid doing so unless you know what you're doing. """ self._sections = {} # Install the sections all kickstart files support. self.registerSection(PreScriptSection(self.handler, dataObj=Script)) self.registerSection(PostScriptSection(self.handler, dataObj=Script)) self.registerSection(TracebackScriptSection(self.handler, dataObj=Script)) self.registerSection(PackageSection(self.handler)) ########NEW FILE######## __FILENAME__ = sections # # sections.py: Kickstart file sections. # # Chris Lumens <clumens@redhat.com> # # Copyright 2011 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # """ This module exports the classes that define a section of a kickstart file. A section is a chunk of the file starting with a %tag and ending with a %end. Examples of sections include %packages, %pre, and %post. You may use this module to define your own custom sections which will be treated just the same as a predefined one by the kickstart parser. All that is necessary is to create a new subclass of Section and call parser.registerSection with an instance of your new class. """ from constants import * from options import KSOptionParser from version import * class Section(object): """The base class for defining kickstart sections. You are free to subclass this as appropriate. Class attributes: allLines -- Does this section require the parser to call handleLine for every line in the section, even blanks and comments? sectionOpen -- The string that denotes the start of this section. You must start your tag with a percent sign. timesSeen -- This attribute is for informational purposes only. It is incremented every time handleHeader is called to keep track of the number of times a section of this type is seen. """ allLines = False sectionOpen = "" timesSeen = 0 def __init__(self, handler, **kwargs): """Create a new Script instance. At the least, you must pass in an instance of a baseHandler subclass. Valid kwargs: dataObj -- """ self.handler = handler self.version = self.handler.version self.dataObj = kwargs.get("dataObj", None) def finalize(self): """This method is called when the %end tag for a section is seen. It is not required to be provided. """ pass def handleLine(self, line): """This method is called for every line of a section. Take whatever action is appropriate. While this method is not required to be provided, not providing it does not make a whole lot of sense. Arguments: line -- The complete line, with any trailing newline. """ pass def handleHeader(self, lineno, args): """This method is called when the opening tag for a section is seen. Not all sections will need this method, though all provided with kickstart include one. Arguments: args -- A list of all strings passed as arguments to the section opening tag. """ self.timesSeen += 1 class NullSection(Section): """This defines a section that pykickstart will recognize but do nothing with. If the parser runs across a %section that has no object registered, it will raise an error. Sometimes, you may want to simply ignore those sections instead. This class is useful for that purpose. """ def __init__(self, *args, **kwargs): """Create a new NullSection instance. You must pass a sectionOpen parameter (including a leading '%') for the section you wish to ignore. """ Section.__init__(self, *args, **kwargs) self.sectionOpen = kwargs.get("sectionOpen") class ScriptSection(Section): allLines = True def __init__(self, *args, **kwargs): Section.__init__(self, *args, **kwargs) self._script = {} self._resetScript() def _getParser(self): op = KSOptionParser(self.version) op.add_option("--erroronfail", dest="errorOnFail", action="store_true", default=False) op.add_option("--interpreter", dest="interpreter", default="/bin/sh") op.add_option("--log", "--logfile", dest="log") return op def _resetScript(self): self._script = {"interp": "/bin/sh", "log": None, "errorOnFail": False, "lineno": None, "chroot": False, "body": []} def handleLine(self, line): self._script["body"].append(line) def finalize(self): if " ".join(self._script["body"]).strip() == "": return kwargs = {"interp": self._script["interp"], "inChroot": self._script["chroot"], "lineno": self._script["lineno"], "logfile": self._script["log"], "errorOnFail": self._script["errorOnFail"], "type": self._script["type"]} s = self.dataObj (self._script["body"], **kwargs) self._resetScript() if self.handler: self.handler.scripts.append(s) def handleHeader(self, lineno, args): """Process the arguments to a %pre/%post/%traceback header for later setting on a Script instance once the end of the script is found. This method may be overridden in a subclass if necessary. """ Section.handleHeader(self, lineno, args) op = self._getParser() (opts, extra) = op.parse_args(args=args[1:], lineno=lineno) self._script["interp"] = opts.interpreter self._script["lineno"] = lineno self._script["log"] = opts.log self._script["errorOnFail"] = opts.errorOnFail if hasattr(opts, "nochroot"): self._script["chroot"] = not opts.nochroot class PreScriptSection(ScriptSection): sectionOpen = "%pre" def _resetScript(self): ScriptSection._resetScript(self) self._script["type"] = KS_SCRIPT_PRE class PostScriptSection(ScriptSection): sectionOpen = "%post" def _getParser(self): op = ScriptSection._getParser(self) op.add_option("--nochroot", dest="nochroot", action="store_true", default=False) return op def _resetScript(self): ScriptSection._resetScript(self) self._script["chroot"] = True self._script["type"] = KS_SCRIPT_POST class TracebackScriptSection(ScriptSection): sectionOpen = "%traceback" def _resetScript(self): ScriptSection._resetScript(self) self._script["type"] = KS_SCRIPT_TRACEBACK class PackageSection(Section): sectionOpen = "%packages" def handleLine(self, line): if not self.handler: return (h, s, t) = line.partition('#') line = h.rstrip() self.handler.packages.add([line]) def handleHeader(self, lineno, args): """Process the arguments to the %packages header and set attributes on the Version's Packages instance appropriate. This method may be overridden in a subclass if necessary. """ Section.handleHeader(self, lineno, args) op = KSOptionParser(version=self.version) op.add_option("--excludedocs", dest="excludedocs", action="store_true", default=False) op.add_option("--ignoremissing", dest="ignoremissing", action="store_true", default=False) op.add_option("--nobase", dest="nobase", action="store_true", default=False) op.add_option("--ignoredeps", dest="resolveDeps", action="store_false", deprecated=FC4, removed=F9) op.add_option("--resolvedeps", dest="resolveDeps", action="store_true", deprecated=FC4, removed=F9) op.add_option("--default", dest="defaultPackages", action="store_true", default=False, introduced=F7) op.add_option("--instLangs", dest="instLangs", type="string", default="", introduced=F9) (opts, extra) = op.parse_args(args=args[1:], lineno=lineno) self.handler.packages.excludeDocs = opts.excludedocs self.handler.packages.addBase = not opts.nobase if opts.ignoremissing: self.handler.packages.handleMissing = KS_MISSING_IGNORE else: self.handler.packages.handleMissing = KS_MISSING_PROMPT if opts.defaultPackages: self.handler.packages.default = True if opts.instLangs: self.handler.packages.instLangs = opts.instLangs ########NEW FILE######## __FILENAME__ = byterange # This library is free software; you can redistribute it and/or # modify it under the terms of the GNU Lesser General Public # License as published by the Free Software Foundation; either # version 2.1 of the License, or (at your option) any later version. # # This library is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # Lesser General Public License for more details. # # You should have received a copy of the GNU Lesser General Public # License along with this library; if not, write to the # Free Software Foundation, Inc., # 59 Temple Place, Suite 330, # Boston, MA 02111-1307 USA # This file is part of urlgrabber, a high-level cross-protocol url-grabber # Copyright 2002-2004 Michael D. Stenner, Ryan Tomayko # $Id: byterange.py,v 1.12 2006/07/20 20:15:58 mstenner Exp $ import os import stat import urllib import urllib2 import rfc822 DEBUG = None try: from cStringIO import StringIO except ImportError, msg: from StringIO import StringIO class RangeError(IOError): """Error raised when an unsatisfiable range is requested.""" pass class HTTPRangeHandler(urllib2.BaseHandler): """Handler that enables HTTP Range headers. This was extremely simple. The Range header is a HTTP feature to begin with so all this class does is tell urllib2 that the "206 Partial Content" reponse from the HTTP server is what we expected. Example: import urllib2 import byterange range_handler = range.HTTPRangeHandler() opener = urllib2.build_opener(range_handler) # install it urllib2.install_opener(opener) # create Request and set Range header req = urllib2.Request('http://www.python.org/') req.header['Range'] = 'bytes=30-50' f = urllib2.urlopen(req) """ def http_error_206(self, req, fp, code, msg, hdrs): # 206 Partial Content Response r = urllib.addinfourl(fp, hdrs, req.get_full_url()) r.code = code r.msg = msg return r def http_error_416(self, req, fp, code, msg, hdrs): # HTTP's Range Not Satisfiable error raise RangeError('Requested Range Not Satisfiable') class HTTPSRangeHandler(HTTPRangeHandler): """ Range Header support for HTTPS. """ def https_error_206(self, req, fp, code, msg, hdrs): return self.http_error_206(req, fp, code, msg, hdrs) def https_error_416(self, req, fp, code, msg, hdrs): self.https_error_416(req, fp, code, msg, hdrs) class RangeableFileObject: """File object wrapper to enable raw range handling. This was implemented primarilary for handling range specifications for file:// urls. This object effectively makes a file object look like it consists only of a range of bytes in the stream. Examples: # expose 10 bytes, starting at byte position 20, from # /etc/aliases. >>> fo = RangeableFileObject(file('/etc/passwd', 'r'), (20,30)) # seek seeks within the range (to position 23 in this case) >>> fo.seek(3) # tell tells where your at _within the range_ (position 3 in # this case) >>> fo.tell() # read EOFs if an attempt is made to read past the last # byte in the range. the following will return only 7 bytes. >>> fo.read(30) """ def __init__(self, fo, rangetup): """Create a RangeableFileObject. fo -- a file like object. only the read() method need be supported but supporting an optimized seek() is preferable. rangetup -- a (firstbyte,lastbyte) tuple specifying the range to work over. The file object provided is assumed to be at byte offset 0. """ self.fo = fo (self.firstbyte, self.lastbyte) = range_tuple_normalize(rangetup) self.realpos = 0 self._do_seek(self.firstbyte) def __getattr__(self, name): """This effectively allows us to wrap at the instance level. Any attribute not found in _this_ object will be searched for in self.fo. This includes methods.""" if hasattr(self.fo, name): return getattr(self.fo, name) raise AttributeError, name def tell(self): """Return the position within the range. This is different from fo.seek in that position 0 is the first byte position of the range tuple. For example, if this object was created with a range tuple of (500,899), tell() will return 0 when at byte position 500 of the file. """ return (self.realpos - self.firstbyte) def seek(self,offset,whence=0): """Seek within the byte range. Positioning is identical to that described under tell(). """ assert whence in (0, 1, 2) if whence == 0: # absolute seek realoffset = self.firstbyte + offset elif whence == 1: # relative seek realoffset = self.realpos + offset elif whence == 2: # absolute from end of file # XXX: are we raising the right Error here? raise IOError('seek from end of file not supported.') # do not allow seek past lastbyte in range if self.lastbyte and (realoffset >= self.lastbyte): realoffset = self.lastbyte self._do_seek(realoffset - self.realpos) def read(self, size=-1): """Read within the range. This method will limit the size read based on the range. """ size = self._calc_read_size(size) rslt = self.fo.read(size) self.realpos += len(rslt) return rslt def readline(self, size=-1): """Read lines within the range. This method will limit the size read based on the range. """ size = self._calc_read_size(size) rslt = self.fo.readline(size) self.realpos += len(rslt) return rslt def _calc_read_size(self, size): """Handles calculating the amount of data to read based on the range. """ if self.lastbyte: if size > -1: if ((self.realpos + size) >= self.lastbyte): size = (self.lastbyte - self.realpos) else: size = (self.lastbyte - self.realpos) return size def _do_seek(self,offset): """Seek based on whether wrapped object supports seek(). offset is relative to the current position (self.realpos). """ assert offset >= 0 if not hasattr(self.fo, 'seek'): self._poor_mans_seek(offset) else: self.fo.seek(self.realpos + offset) self.realpos+= offset def _poor_mans_seek(self,offset): """Seek by calling the wrapped file objects read() method. This is used for file like objects that do not have native seek support. The wrapped objects read() method is called to manually seek to the desired position. offset -- read this number of bytes from the wrapped file object. raise RangeError if we encounter EOF before reaching the specified offset. """ pos = 0 bufsize = 1024 while pos < offset: if (pos + bufsize) > offset: bufsize = offset - pos buf = self.fo.read(bufsize) if len(buf) != bufsize: raise RangeError('Requested Range Not Satisfiable') pos+= bufsize class FileRangeHandler(urllib2.FileHandler): """FileHandler subclass that adds Range support. This class handles Range headers exactly like an HTTP server would. """ def open_local_file(self, req): import mimetypes import mimetools host = req.get_host() file = req.get_selector() localfile = urllib.url2pathname(file) stats = os.stat(localfile) size = stats[stat.ST_SIZE] modified = rfc822.formatdate(stats[stat.ST_MTIME]) mtype = mimetypes.guess_type(file)[0] if host: host, port = urllib.splitport(host) if port or socket.gethostbyname(host) not in self.get_names(): raise urllib2.URLError('file not on local host') fo = open(localfile,'rb') brange = req.headers.get('Range',None) brange = range_header_to_tuple(brange) assert brange != () if brange: (fb,lb) = brange if lb == '': lb = size if fb < 0 or fb > size or lb > size: raise RangeError('Requested Range Not Satisfiable') size = (lb - fb) fo = RangeableFileObject(fo, (fb,lb)) headers = mimetools.Message(StringIO( 'Content-Type: %s\nContent-Length: %d\nLast-modified: %s\n' % (mtype or 'text/plain', size, modified))) return urllib.addinfourl(fo, headers, 'file:'+file) # FTP Range Support # Unfortunately, a large amount of base FTP code had to be copied # from urllib and urllib2 in order to insert the FTP REST command. # Code modifications for range support have been commented as # follows: # -- range support modifications start/end here from urllib import splitport, splituser, splitpasswd, splitattr, \ unquote, addclosehook, addinfourl import ftplib import socket import sys import ftplib import mimetypes import mimetools class FTPRangeHandler(urllib2.FTPHandler): def ftp_open(self, req): host = req.get_host() if not host: raise IOError, ('ftp error', 'no host given') host, port = splitport(host) if port is None: port = ftplib.FTP_PORT # username/password handling user, host = splituser(host) if user: user, passwd = splitpasswd(user) else: passwd = None host = unquote(host) user = unquote(user or '') passwd = unquote(passwd or '') try: host = socket.gethostbyname(host) except socket.error, msg: raise urllib2.URLError(msg) path, attrs = splitattr(req.get_selector()) dirs = path.split('/') dirs = map(unquote, dirs) dirs, file = dirs[:-1], dirs[-1] if dirs and not dirs[0]: dirs = dirs[1:] try: fw = self.connect_ftp(user, passwd, host, port, dirs) type = file and 'I' or 'D' for attr in attrs: attr, value = splitattr(attr) if attr.lower() == 'type' and \ value in ('a', 'A', 'i', 'I', 'd', 'D'): type = value.upper() # -- range support modifications start here rest = None range_tup = range_header_to_tuple(req.headers.get('Range',None)) assert range_tup != () if range_tup: (fb,lb) = range_tup if fb > 0: rest = fb # -- range support modifications end here fp, retrlen = fw.retrfile(file, type, rest) # -- range support modifications start here if range_tup: (fb,lb) = range_tup if lb == '': if retrlen is None or retrlen == 0: raise RangeError('Requested Range Not Satisfiable due to unobtainable file length.') lb = retrlen retrlen = lb - fb if retrlen < 0: # beginning of range is larger than file raise RangeError('Requested Range Not Satisfiable') else: retrlen = lb - fb fp = RangeableFileObject(fp, (0,retrlen)) # -- range support modifications end here headers = "" mtype = mimetypes.guess_type(req.get_full_url())[0] if mtype: headers += "Content-Type: %s\n" % mtype if retrlen is not None and retrlen >= 0: headers += "Content-Length: %d\n" % retrlen sf = StringIO(headers) headers = mimetools.Message(sf) return addinfourl(fp, headers, req.get_full_url()) except ftplib.all_errors, msg: raise IOError, ('ftp error', msg), sys.exc_info()[2] def connect_ftp(self, user, passwd, host, port, dirs): fw = ftpwrapper(user, passwd, host, port, dirs) return fw class ftpwrapper(urllib.ftpwrapper): # range support note: # this ftpwrapper code is copied directly from # urllib. The only enhancement is to add the rest # argument and pass it on to ftp.ntransfercmd def retrfile(self, file, type, rest=None): self.endtransfer() if type in ('d', 'D'): cmd = 'TYPE A'; isdir = 1 else: cmd = 'TYPE ' + type; isdir = 0 try: self.ftp.voidcmd(cmd) except ftplib.all_errors: self.init() self.ftp.voidcmd(cmd) conn = None if file and not isdir: # Use nlst to see if the file exists at all try: self.ftp.nlst(file) except ftplib.error_perm, reason: raise IOError, ('ftp error', reason), sys.exc_info()[2] # Restore the transfer mode! self.ftp.voidcmd(cmd) # Try to retrieve as a file try: cmd = 'RETR ' + file conn = self.ftp.ntransfercmd(cmd, rest) except ftplib.error_perm, reason: if str(reason)[:3] == '501': # workaround for REST not supported error fp, retrlen = self.retrfile(file, type) fp = RangeableFileObject(fp, (rest,'')) return (fp, retrlen) elif str(reason)[:3] != '550': raise IOError, ('ftp error', reason), sys.exc_info()[2] if not conn: # Set transfer mode to ASCII! self.ftp.voidcmd('TYPE A') # Try a directory listing if file: cmd = 'LIST ' + file else: cmd = 'LIST' conn = self.ftp.ntransfercmd(cmd) self.busy = 1 # Pass back both a suitably decorated object and a retrieval length return (addclosehook(conn[0].makefile('rb'), self.endtransfer), conn[1]) #################################################################### # Range Tuple Functions # XXX: These range tuple functions might go better in a class. _rangere = None def range_header_to_tuple(range_header): """Get a (firstbyte,lastbyte) tuple from a Range header value. Range headers have the form "bytes=<firstbyte>-<lastbyte>". This function pulls the firstbyte and lastbyte values and returns a (firstbyte,lastbyte) tuple. If lastbyte is not specified in the header value, it is returned as an empty string in the tuple. Return None if range_header is None Return () if range_header does not conform to the range spec pattern. """ global _rangere if range_header is None: return None if _rangere is None: import re _rangere = re.compile(r'^bytes=(\d{1,})-(\d*)') match = _rangere.match(range_header) if match: tup = range_tuple_normalize(match.group(1,2)) if tup and tup[1]: tup = (tup[0],tup[1]+1) return tup return () def range_tuple_to_header(range_tup): """Convert a range tuple to a Range header value. Return a string of the form "bytes=<firstbyte>-<lastbyte>" or None if no range is needed. """ if range_tup is None: return None range_tup = range_tuple_normalize(range_tup) if range_tup: if range_tup[1]: range_tup = (range_tup[0],range_tup[1] - 1) return 'bytes=%s-%s' % range_tup def range_tuple_normalize(range_tup): """Normalize a (first_byte,last_byte) range tuple. Return a tuple whose first element is guaranteed to be an int and whose second element will be '' (meaning: the last byte) or an int. Finally, return None if the normalized tuple == (0,'') as that is equivelant to retrieving the entire file. """ if range_tup is None: return None # handle first byte fb = range_tup[0] if fb in (None,''): fb = 0 else: fb = int(fb) # handle last byte try: lb = range_tup[1] except IndexError: lb = '' else: if lb is None: lb = '' elif lb != '': lb = int(lb) # check if range is over the entire file if (fb,lb) == (0,''): return None # check that the range is valid if lb < fb: raise RangeError('Invalid byte range: %s-%s' % (fb,lb)) return (fb,lb) ########NEW FILE######## __FILENAME__ = grabber # This library is free software; you can redistribute it and/or # modify it under the terms of the GNU Lesser General Public # License as published by the Free Software Foundation; either # version 2.1 of the License, or (at your option) any later version. # # This library is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # Lesser General Public License for more details. # # You should have received a copy of the GNU Lesser General Public # License along with this library; if not, write to the # Free Software Foundation, Inc., # 59 Temple Place, Suite 330, # Boston, MA 02111-1307 USA # This file is part of urlgrabber, a high-level cross-protocol url-grabber # Copyright 2002-2004 Michael D. Stenner, Ryan Tomayko """A high-level cross-protocol url-grabber. GENERAL ARGUMENTS (kwargs) Where possible, the module-level default is indicated, and legal values are provided. copy_local = 0 [0|1] ignored except for file:// urls, in which case it specifies whether urlgrab should still make a copy of the file, or simply point to the existing copy. The module level default for this option is 0. close_connection = 0 [0|1] tells URLGrabber to close the connection after a file has been transfered. This is ignored unless the download happens with the http keepalive handler (keepalive=1). Otherwise, the connection is left open for further use. The module level default for this option is 0 (keepalive connections will not be closed). keepalive = 1 [0|1] specifies whether keepalive should be used for HTTP/1.1 servers that support it. The module level default for this option is 1 (keepalive is enabled). progress_obj = None a class instance that supports the following methods: po.start(filename, url, basename, length, text) # length will be None if unknown po.update(read) # read == bytes read so far po.end() text = None specifies an alternativ text item in the beginning of the progress bar line. If not given, the basename of the file is used. throttle = 1.0 a number - if it's an int, it's the bytes/second throttle limit. If it's a float, it is first multiplied by bandwidth. If throttle == 0, throttling is disabled. If None, the module-level default (which can be set on default_grabber.throttle) is used. See BANDWIDTH THROTTLING for more information. timeout = None a positive float expressing the number of seconds to wait for socket operations. If the value is None or 0.0, socket operations will block forever. Setting this option causes urlgrabber to call the settimeout method on the Socket object used for the request. See the Python documentation on settimeout for more information. http://www.python.org/doc/current/lib/socket-objects.html bandwidth = 0 the nominal max bandwidth in bytes/second. If throttle is a float and bandwidth == 0, throttling is disabled. If None, the module-level default (which can be set on default_grabber.bandwidth) is used. See BANDWIDTH THROTTLING for more information. range = None a tuple of the form (first_byte, last_byte) describing a byte range to retrieve. Either or both of the values may set to None. If first_byte is None, byte offset 0 is assumed. If last_byte is None, the last byte available is assumed. Note that the range specification is python-like in that (0,10) will yeild the first 10 bytes of the file. If set to None, no range will be used. reget = None [None|'simple'|'check_timestamp'] whether to attempt to reget a partially-downloaded file. Reget only applies to .urlgrab and (obviously) only if there is a partially downloaded file. Reget has two modes: 'simple' -- the local file will always be trusted. If there are 100 bytes in the local file, then the download will always begin 100 bytes into the requested file. 'check_timestamp' -- the timestamp of the server file will be compared to the timestamp of the local file. ONLY if the local file is newer than or the same age as the server file will reget be used. If the server file is newer, or the timestamp is not returned, the entire file will be fetched. NOTE: urlgrabber can do very little to verify that the partial file on disk is identical to the beginning of the remote file. You may want to either employ a custom "checkfunc" or simply avoid using reget in situations where corruption is a concern. user_agent = 'urlgrabber/VERSION' a string, usually of the form 'AGENT/VERSION' that is provided to HTTP servers in the User-agent header. The module level default for this option is "urlgrabber/VERSION". http_headers = None a tuple of 2-tuples, each containing a header and value. These will be used for http and https requests only. For example, you can do http_headers = (('Pragma', 'no-cache'),) ftp_headers = None this is just like http_headers, but will be used for ftp requests. proxies = None a dictionary that maps protocol schemes to proxy hosts. For example, to use a proxy server on host "foo" port 3128 for http and https URLs: proxies={ 'http' : 'http://foo:3128', 'https' : 'http://foo:3128' } note that proxy authentication information may be provided using normal URL constructs: proxies={ 'http' : 'http://user:host@foo:3128' } Lastly, if proxies is None, the default environment settings will be used. prefix = None a url prefix that will be prepended to all requested urls. For example: g = URLGrabber(prefix='http://foo.com/mirror/') g.urlgrab('some/file.txt') ## this will fetch 'http://foo.com/mirror/some/file.txt' This option exists primarily to allow identical behavior to MirrorGroup (and derived) instances. Note: a '/' will be inserted if necessary, so you cannot specify a prefix that ends with a partial file or directory name. opener = None Overrides the default urllib2.OpenerDirector provided to urllib2 when making requests. This option exists so that the urllib2 handler chain may be customized. Note that the range, reget, proxy, and keepalive features require that custom handlers be provided to urllib2 in order to function properly. If an opener option is provided, no attempt is made by urlgrabber to ensure chain integrity. You are responsible for ensuring that any extension handlers are present if said features are required. data = None Only relevant for the HTTP family (and ignored for other protocols), this allows HTTP POSTs. When the data kwarg is present (and not None), an HTTP request will automatically become a POST rather than GET. This is done by direct passthrough to urllib2. If you use this, you may also want to set the 'Content-length' and 'Content-type' headers with the http_headers option. Note that python 2.2 handles the case of these badly and if you do not use the proper case (shown here), your values will be overridden with the defaults. RETRY RELATED ARGUMENTS retry = None the number of times to retry the grab before bailing. If this is zero, it will retry forever. This was intentional... really, it was :). If this value is not supplied or is supplied but is None retrying does not occur. retrycodes = [-1,2,4,5,6,7] a sequence of errorcodes (values of e.errno) for which it should retry. See the doc on URLGrabError for more details on this. You might consider modifying a copy of the default codes rather than building yours from scratch so that if the list is extended in the future (or one code is split into two) you can still enjoy the benefits of the default list. You can do that with something like this: retrycodes = urlgrabber.grabber.URLGrabberOptions().retrycodes if 12 not in retrycodes: retrycodes.append(12) checkfunc = None a function to do additional checks. This defaults to None, which means no additional checking. The function should simply return on a successful check. It should raise URLGrabError on an unsuccessful check. Raising of any other exception will be considered immediate failure and no retries will occur. If it raises URLGrabError, the error code will determine the retry behavior. Negative error numbers are reserved for use by these passed in functions, so you can use many negative numbers for different types of failure. By default, -1 results in a retry, but this can be customized with retrycodes. If you simply pass in a function, it will be given exactly one argument: a CallbackObject instance with the .url attribute defined and either .filename (for urlgrab) or .data (for urlread). For urlgrab, .filename is the name of the local file. For urlread, .data is the actual string data. If you need other arguments passed to the callback (program state of some sort), you can do so like this: checkfunc=(function, ('arg1', 2), {'kwarg': 3}) if the downloaded file has filename /tmp/stuff, then this will result in this call (for urlgrab): function(obj, 'arg1', 2, kwarg=3) # obj.filename = '/tmp/stuff' # obj.url = 'http://foo.com/stuff' NOTE: both the "args" tuple and "kwargs" dict must be present if you use this syntax, but either (or both) can be empty. failure_callback = None The callback that gets called during retries when an attempt to fetch a file fails. The syntax for specifying the callback is identical to checkfunc, except for the attributes defined in the CallbackObject instance. The attributes for failure_callback are: exception = the raised exception url = the url we're trying to fetch tries = the number of tries so far (including this one) retry = the value of the retry option The callback is present primarily to inform the calling program of the failure, but if it raises an exception (including the one it's passed) that exception will NOT be caught and will therefore cause future retries to be aborted. The callback is called for EVERY failure, including the last one. On the last try, the callback can raise an alternate exception, but it cannot (without severe trickiness) prevent the exception from being raised. interrupt_callback = None This callback is called if KeyboardInterrupt is received at any point in the transfer. Basically, this callback can have three impacts on the fetch process based on the way it exits: 1) raise no exception: the current fetch will be aborted, but any further retries will still take place 2) raise a URLGrabError: if you're using a MirrorGroup, then this will prompt a failover to the next mirror according to the behavior of the MirrorGroup subclass. It is recommended that you raise URLGrabError with code 15, 'user abort'. If you are NOT using a MirrorGroup subclass, then this is the same as (3). 3) raise some other exception (such as KeyboardInterrupt), which will not be caught at either the grabber or mirror levels. That is, it will be raised up all the way to the caller. This callback is very similar to failure_callback. They are passed the same arguments, so you could use the same function for both. urlparser = URLParser() The URLParser class handles pre-processing of URLs, including auth-handling for user/pass encoded in http urls, file handing (that is, filenames not sent as a URL), and URL quoting. If you want to override any of this behavior, you can pass in a replacement instance. See also the 'quote' option. quote = None Whether or not to quote the path portion of a url. quote = 1 -> quote the URLs (they're not quoted yet) quote = 0 -> do not quote them (they're already quoted) quote = None -> guess what to do This option only affects proper urls like 'file:///etc/passwd'; it does not affect 'raw' filenames like '/etc/passwd'. The latter will always be quoted as they are converted to URLs. Also, only the path part of a url is quoted. If you need more fine-grained control, you should probably subclass URLParser and pass it in via the 'urlparser' option. BANDWIDTH THROTTLING urlgrabber supports throttling via two values: throttle and bandwidth Between the two, you can either specify and absolute throttle threshold or specify a theshold as a fraction of maximum available bandwidth. throttle is a number - if it's an int, it's the bytes/second throttle limit. If it's a float, it is first multiplied by bandwidth. If throttle == 0, throttling is disabled. If None, the module-level default (which can be set with set_throttle) is used. bandwidth is the nominal max bandwidth in bytes/second. If throttle is a float and bandwidth == 0, throttling is disabled. If None, the module-level default (which can be set with set_bandwidth) is used. THROTTLING EXAMPLES: Lets say you have a 100 Mbps connection. This is (about) 10^8 bits per second, or 12,500,000 Bytes per second. You have a number of throttling options: *) set_bandwidth(12500000); set_throttle(0.5) # throttle is a float This will limit urlgrab to use half of your available bandwidth. *) set_throttle(6250000) # throttle is an int This will also limit urlgrab to use half of your available bandwidth, regardless of what bandwidth is set to. *) set_throttle(6250000); set_throttle(1.0) # float Use half your bandwidth *) set_throttle(6250000); set_throttle(2.0) # float Use up to 12,500,000 Bytes per second (your nominal max bandwidth) *) set_throttle(6250000); set_throttle(0) # throttle = 0 Disable throttling - this is more efficient than a very large throttle setting. *) set_throttle(0); set_throttle(1.0) # throttle is float, bandwidth = 0 Disable throttling - this is the default when the module is loaded. SUGGESTED AUTHOR IMPLEMENTATION (THROTTLING) While this is flexible, it's not extremely obvious to the user. I suggest you implement a float throttle as a percent to make the distinction between absolute and relative throttling very explicit. Also, you may want to convert the units to something more convenient than bytes/second, such as kbps or kB/s, etc. """ # $Id: grabber.py,v 1.48 2006/09/22 00:58:05 mstenner Exp $ import os import os.path import sys import urlparse import rfc822 import time import string import urllib import urllib2 from stat import * # S_* and ST_* ######################################################################## # MODULE INITIALIZATION ######################################################################## try: exec('from ' + (__name__.split('.'))[0] + ' import __version__') except: __version__ = '???' import sslfactory auth_handler = urllib2.HTTPBasicAuthHandler( \ urllib2.HTTPPasswordMgrWithDefaultRealm()) try: from i18n import _ except ImportError, msg: def _(st): return st try: from httplib import HTTPException except ImportError, msg: HTTPException = None try: # This is a convenient way to make keepalive optional. # Just rename the module so it can't be imported. import keepalive from keepalive import HTTPHandler, HTTPSHandler have_keepalive = True except ImportError, msg: have_keepalive = False try: # add in range support conditionally too import byterange from byterange import HTTPRangeHandler, HTTPSRangeHandler, \ FileRangeHandler, FTPRangeHandler, range_tuple_normalize, \ range_tuple_to_header, RangeError except ImportError, msg: range_handlers = () RangeError = None have_range = 0 else: range_handlers = (HTTPRangeHandler(), HTTPSRangeHandler(), FileRangeHandler(), FTPRangeHandler()) have_range = 1 # check whether socket timeout support is available (Python >= 2.3) import socket try: TimeoutError = socket.timeout have_socket_timeout = True except AttributeError: TimeoutError = None have_socket_timeout = False ######################################################################## # functions for debugging output. These functions are here because they # are also part of the module initialization. DEBUG = None def set_logger(DBOBJ): """Set the DEBUG object. This is called by _init_default_logger when the environment variable URLGRABBER_DEBUG is set, but can also be called by a calling program. Basically, if the calling program uses the logging module and would like to incorporate urlgrabber logging, then it can do so this way. It's probably not necessary as most internal logging is only for debugging purposes. The passed-in object should be a logging.Logger instance. It will be pushed into the keepalive and byterange modules if they're being used. The mirror module pulls this object in on import, so you will need to manually push into it. In fact, you may find it tidier to simply push your logging object (or objects) into each of these modules independently. """ global DEBUG DEBUG = DBOBJ if have_keepalive and keepalive.DEBUG is None: keepalive.DEBUG = DBOBJ if have_range and byterange.DEBUG is None: byterange.DEBUG = DBOBJ if sslfactory.DEBUG is None: sslfactory.DEBUG = DBOBJ def _init_default_logger(): '''Examines the environment variable URLGRABBER_DEBUG and creates a logging object (logging.logger) based on the contents. It takes the form URLGRABBER_DEBUG=level,filename where "level" can be either an integer or a log level from the logging module (DEBUG, INFO, etc). If the integer is zero or less, logging will be disabled. Filename is the filename where logs will be sent. If it is "-", then stdout will be used. If the filename is empty or missing, stderr will be used. If the variable cannot be processed or the logging module cannot be imported (python < 2.3) then logging will be disabled. Here are some examples: URLGRABBER_DEBUG=1,debug.txt # log everything to debug.txt URLGRABBER_DEBUG=WARNING,- # log warning and higher to stdout URLGRABBER_DEBUG=INFO # log info and higher to stderr This funtion is called during module initialization. It is not intended to be called from outside. The only reason it is a function at all is to keep the module-level namespace tidy and to collect the code into a nice block.''' try: dbinfo = os.environ['URLGRABBER_DEBUG'].split(',') import logging level = logging._levelNames.get(dbinfo[0], int(dbinfo[0])) if level < 1: raise ValueError() formatter = logging.Formatter('%(asctime)s %(message)s') if len(dbinfo) > 1: filename = dbinfo[1] else: filename = '' if filename == '': handler = logging.StreamHandler(sys.stderr) elif filename == '-': handler = logging.StreamHandler(sys.stdout) else: handler = logging.FileHandler(filename) handler.setFormatter(formatter) DBOBJ = logging.getLogger('urlgrabber') DBOBJ.addHandler(handler) DBOBJ.setLevel(level) except (KeyError, ImportError, ValueError): DBOBJ = None set_logger(DBOBJ) _init_default_logger() ######################################################################## # END MODULE INITIALIZATION ######################################################################## class URLGrabError(IOError): """ URLGrabError error codes: URLGrabber error codes (0 -- 255) 0 - everything looks good (you should never see this) 1 - malformed url 2 - local file doesn't exist 3 - request for non-file local file (dir, etc) 4 - IOError on fetch 5 - OSError on fetch 6 - no content length header when we expected one 7 - HTTPException 8 - Exceeded read limit (for urlread) 9 - Requested byte range not satisfiable. 10 - Byte range requested, but range support unavailable 11 - Illegal reget mode 12 - Socket timeout 13 - malformed proxy url 14 - HTTPError (includes .code and .exception attributes) 15 - user abort MirrorGroup error codes (256 -- 511) 256 - No more mirrors left to try Custom (non-builtin) classes derived from MirrorGroup (512 -- 767) [ this range reserved for application-specific error codes ] Retry codes (< 0) -1 - retry the download, unknown reason Note: to test which group a code is in, you can simply do integer division by 256: e.errno / 256 Negative codes are reserved for use by functions passed in to retrygrab with checkfunc. The value -1 is built in as a generic retry code and is already included in the retrycodes list. Therefore, you can create a custom check function that simply returns -1 and the fetch will be re-tried. For more customized retries, you can use other negative number and include them in retry-codes. This is nice for outputting useful messages about what failed. You can use these error codes like so: try: urlgrab(url) except URLGrabError, e: if e.errno == 3: ... # or print e.strerror # or simply print e #### print '[Errno %i] %s' % (e.errno, e.strerror) """ pass class CallbackObject: """Container for returned callback data. This is currently a dummy class into which urlgrabber can stuff information for passing to callbacks. This way, the prototype for all callbacks is the same, regardless of the data that will be passed back. Any function that accepts a callback function as an argument SHOULD document what it will define in this object. It is possible that this class will have some greater functionality in the future. """ def __init__(self, **kwargs): self.__dict__.update(kwargs) def urlgrab(url, filename=None, **kwargs): """grab the file at <url> and make a local copy at <filename> If filename is none, the basename of the url is used. urlgrab returns the filename of the local file, which may be different from the passed-in filename if the copy_local kwarg == 0. See module documentation for a description of possible kwargs. """ return default_grabber.urlgrab(url, filename, **kwargs) def urlopen(url, **kwargs): """open the url and return a file object If a progress object or throttle specifications exist, then a special file object will be returned that supports them. The file object can be treated like any other file object. See module documentation for a description of possible kwargs. """ return default_grabber.urlopen(url, **kwargs) def urlread(url, limit=None, **kwargs): """read the url into a string, up to 'limit' bytes If the limit is exceeded, an exception will be thrown. Note that urlread is NOT intended to be used as a way of saying "I want the first N bytes" but rather 'read the whole file into memory, but don't use too much' See module documentation for a description of possible kwargs. """ return default_grabber.urlread(url, limit, **kwargs) class URLParser: """Process the URLs before passing them to urllib2. This class does several things: * add any prefix * translate a "raw" file to a proper file: url * handle any http or https auth that's encoded within the url * quote the url Only the "parse" method is called directly, and it calls sub-methods. An instance of this class is held in the options object, which means that it's easy to change the behavior by sub-classing and passing the replacement in. It need only have a method like: url, parts = urlparser.parse(url, opts) """ def parse(self, url, opts): """parse the url and return the (modified) url and its parts Note: a raw file WILL be quoted when it's converted to a URL. However, other urls (ones which come with a proper scheme) may or may not be quoted according to opts.quote opts.quote = 1 --> quote it opts.quote = 0 --> do not quote it opts.quote = None --> guess """ quote = opts.quote if opts.prefix: url = self.add_prefix(url, opts.prefix) parts = urlparse.urlparse(url) (scheme, host, path, parm, query, frag) = parts if not scheme or (len(scheme) == 1 and scheme in string.letters): # if a scheme isn't specified, we guess that it's "file:" if url[0] not in '/\\': url = os.path.abspath(url) url = 'file:' + urllib.pathname2url(url) parts = urlparse.urlparse(url) quote = 0 # pathname2url quotes, so we won't do it again if scheme in ['http', 'https']: parts = self.process_http(parts) if quote is None: quote = self.guess_should_quote(parts) if quote: parts = self.quote(parts) url = urlparse.urlunparse(parts) return url, parts def add_prefix(self, url, prefix): if prefix[-1] == '/' or url[0] == '/': url = prefix + url else: url = prefix + '/' + url return url def process_http(self, parts): (scheme, host, path, parm, query, frag) = parts if '@' in host and auth_handler: try: user_pass, host = host.split('@', 1) if ':' in user_pass: user, password = user_pass.split(':', 1) except ValueError, e: raise URLGrabError(1, _('Bad URL: %s') % url) if DEBUG: DEBUG.info('adding HTTP auth: %s, XXXXXXXX', user) auth_handler.add_password(None, host, user, password) return (scheme, host, path, parm, query, frag) def quote(self, parts): """quote the URL This method quotes ONLY the path part. If you need to quote other parts, you should override this and pass in your derived class. The other alternative is to quote other parts before passing into urlgrabber. """ (scheme, host, path, parm, query, frag) = parts path = urllib.quote(path) return (scheme, host, path, parm, query, frag) hexvals = '0123456789ABCDEF' def guess_should_quote(self, parts): """ Guess whether we should quote a path. This amounts to guessing whether it's already quoted. find ' ' -> 1 find '%' -> 1 find '%XX' -> 0 else -> 1 """ (scheme, host, path, parm, query, frag) = parts if ' ' in path: return 1 ind = string.find(path, '%') if ind > -1: while ind > -1: if len(path) < ind+3: return 1 code = path[ind+1:ind+3].upper() if code[0] not in self.hexvals or \ code[1] not in self.hexvals: return 1 ind = string.find(path, '%', ind+1) return 0 return 1 class URLGrabberOptions: """Class to ease kwargs handling.""" def __init__(self, delegate=None, **kwargs): """Initialize URLGrabberOptions object. Set default values for all options and then update options specified in kwargs. """ self.delegate = delegate if delegate is None: self._set_defaults() self._set_attributes(**kwargs) def __getattr__(self, name): if self.delegate and hasattr(self.delegate, name): return getattr(self.delegate, name) raise AttributeError, name def raw_throttle(self): """Calculate raw throttle value from throttle and bandwidth values. """ if self.throttle <= 0: return 0 elif type(self.throttle) == type(0): return float(self.throttle) else: # throttle is a float return self.bandwidth * self.throttle def derive(self, **kwargs): """Create a derived URLGrabberOptions instance. This method creates a new instance and overrides the options specified in kwargs. """ return URLGrabberOptions(delegate=self, **kwargs) def _set_attributes(self, **kwargs): """Update object attributes with those provided in kwargs.""" self.__dict__.update(kwargs) if have_range and kwargs.has_key('range'): # normalize the supplied range value self.range = range_tuple_normalize(self.range) if not self.reget in [None, 'simple', 'check_timestamp']: raise URLGrabError(11, _('Illegal reget mode: %s') \ % (self.reget, )) def _set_defaults(self): """Set all options to their default values. When adding new options, make sure a default is provided here. """ self.progress_obj = None self.throttle = 1.0 self.bandwidth = 0 self.retry = None self.retrycodes = [-1,2,4,5,6,7] self.checkfunc = None self.copy_local = 0 self.close_connection = 0 self.range = None self.user_agent = 'urlgrabber/%s' % __version__ self.keepalive = 1 self.proxies = None self.reget = None self.failure_callback = None self.interrupt_callback = None self.prefix = None self.opener = None self.cache_openers = True self.timeout = None self.text = None self.http_headers = None self.ftp_headers = None self.data = None self.urlparser = URLParser() self.quote = None self.ssl_ca_cert = None self.ssl_context = None class URLGrabber: """Provides easy opening of URLs with a variety of options. All options are specified as kwargs. Options may be specified when the class is created and may be overridden on a per request basis. New objects inherit default values from default_grabber. """ def __init__(self, **kwargs): self.opts = URLGrabberOptions(**kwargs) def _retry(self, opts, func, *args): tries = 0 while 1: # there are only two ways out of this loop. The second has # several "sub-ways" # 1) via the return in the "try" block # 2) by some exception being raised # a) an excepton is raised that we don't "except" # b) a callback raises ANY exception # c) we're not retry-ing or have run out of retries # d) the URLGrabError code is not in retrycodes # beware of infinite loops :) tries = tries + 1 exception = None retrycode = None callback = None if DEBUG: DEBUG.info('attempt %i/%s: %s', tries, opts.retry, args[0]) try: r = apply(func, (opts,) + args, {}) if DEBUG: DEBUG.info('success') return r except URLGrabError, e: exception = e callback = opts.failure_callback retrycode = e.errno except KeyboardInterrupt, e: exception = e callback = opts.interrupt_callback if DEBUG: DEBUG.info('exception: %s', exception) if callback: if DEBUG: DEBUG.info('calling callback: %s', callback) cb_func, cb_args, cb_kwargs = self._make_callback(callback) obj = CallbackObject(exception=exception, url=args[0], tries=tries, retry=opts.retry) cb_func(obj, *cb_args, **cb_kwargs) if (opts.retry is None) or (tries == opts.retry): if DEBUG: DEBUG.info('retries exceeded, re-raising') raise if (retrycode is not None) and (retrycode not in opts.retrycodes): if DEBUG: DEBUG.info('retrycode (%i) not in list %s, re-raising', retrycode, opts.retrycodes) raise def urlopen(self, url, **kwargs): """open the url and return a file object If a progress object or throttle value specified when this object was created, then a special file object will be returned that supports them. The file object can be treated like any other file object. """ opts = self.opts.derive(**kwargs) (url,parts) = opts.urlparser.parse(url, opts) def retryfunc(opts, url): return URLGrabberFileObject(url, filename=None, opts=opts) return self._retry(opts, retryfunc, url) def urlgrab(self, url, filename=None, **kwargs): """grab the file at <url> and make a local copy at <filename> If filename is none, the basename of the url is used. urlgrab returns the filename of the local file, which may be different from the passed-in filename if copy_local == 0. """ opts = self.opts.derive(**kwargs) (url,parts) = opts.urlparser.parse(url, opts) (scheme, host, path, parm, query, frag) = parts if filename is None: filename = os.path.basename( urllib.unquote(path) ) if scheme == 'file' and not opts.copy_local: # just return the name of the local file - don't make a # copy currently path = urllib.url2pathname(path) if host: path = os.path.normpath('//' + host + path) if not os.path.exists(path): raise URLGrabError(2, _('Local file does not exist: %s') % (path, )) elif not os.path.isfile(path): raise URLGrabError(3, _('Not a normal file: %s') % (path, )) elif not opts.range: return path def retryfunc(opts, url, filename): fo = URLGrabberFileObject(url, filename, opts) try: fo._do_grab() if not opts.checkfunc is None: cb_func, cb_args, cb_kwargs = \ self._make_callback(opts.checkfunc) obj = CallbackObject() obj.filename = filename obj.url = url apply(cb_func, (obj, )+cb_args, cb_kwargs) finally: fo.close() return filename return self._retry(opts, retryfunc, url, filename) def urlread(self, url, limit=None, **kwargs): """read the url into a string, up to 'limit' bytes If the limit is exceeded, an exception will be thrown. Note that urlread is NOT intended to be used as a way of saying "I want the first N bytes" but rather 'read the whole file into memory, but don't use too much' """ opts = self.opts.derive(**kwargs) (url,parts) = opts.urlparser.parse(url, opts) if limit is not None: limit = limit + 1 def retryfunc(opts, url, limit): fo = URLGrabberFileObject(url, filename=None, opts=opts) s = '' try: # this is an unfortunate thing. Some file-like objects # have a default "limit" of None, while the built-in (real) # file objects have -1. They each break the other, so for # now, we just force the default if necessary. if limit is None: s = fo.read() else: s = fo.read(limit) if not opts.checkfunc is None: cb_func, cb_args, cb_kwargs = \ self._make_callback(opts.checkfunc) obj = CallbackObject() obj.data = s obj.url = url apply(cb_func, (obj, )+cb_args, cb_kwargs) finally: fo.close() return s s = self._retry(opts, retryfunc, url, limit) if limit and len(s) > limit: raise URLGrabError(8, _('Exceeded limit (%i): %s') % (limit, url)) return s def _make_callback(self, callback_obj): if callable(callback_obj): return callback_obj, (), {} else: return callback_obj # create the default URLGrabber used by urlXXX functions. # NOTE: actual defaults are set in URLGrabberOptions default_grabber = URLGrabber() class URLGrabberFileObject: """This is a file-object wrapper that supports progress objects and throttling. This exists to solve the following problem: lets say you want to drop-in replace a normal open with urlopen. You want to use a progress meter and/or throttling, but how do you do that without rewriting your code? Answer: urlopen will return a wrapped file object that does the progress meter and-or throttling internally. """ def __init__(self, url, filename, opts): self.url = url self.filename = filename self.opts = opts self.fo = None self._rbuf = '' self._rbufsize = 1024*8 self._ttime = time.time() self._tsize = 0 self._amount_read = 0 self._opener = None self._do_open() def __getattr__(self, name): """This effectively allows us to wrap at the instance level. Any attribute not found in _this_ object will be searched for in self.fo. This includes methods.""" if hasattr(self.fo, name): return getattr(self.fo, name) raise AttributeError, name def _get_opener(self): """Build a urllib2 OpenerDirector based on request options.""" if self.opts.opener: return self.opts.opener elif self._opener is None: handlers = [] need_keepalive_handler = (have_keepalive and self.opts.keepalive) need_range_handler = (range_handlers and \ (self.opts.range or self.opts.reget)) # if you specify a ProxyHandler when creating the opener # it _must_ come before all other handlers in the list or urllib2 # chokes. if self.opts.proxies: handlers.append( CachedProxyHandler(self.opts.proxies) ) # ------------------------------------------------------- # OK, these next few lines are a serious kludge to get # around what I think is a bug in python 2.2's # urllib2. The basic idea is that default handlers # get applied first. If you override one (like a # proxy handler), then the default gets pulled, but # the replacement goes on the end. In the case of # proxies, this means the normal handler picks it up # first and the proxy isn't used. Now, this probably # only happened with ftp or non-keepalive http, so not # many folks saw it. The simple approach to fixing it # is just to make sure you override the other # conflicting defaults as well. I would LOVE to see # these go way or be dealt with more elegantly. The # problem isn't there after 2.2. -MDS 2005/02/24 if not need_keepalive_handler: handlers.append( urllib2.HTTPHandler() ) if not need_range_handler: handlers.append( urllib2.FTPHandler() ) # ------------------------------------------------------- ssl_factory = sslfactory.get_factory(self.opts.ssl_ca_cert, self.opts.ssl_context) if need_keepalive_handler: handlers.append(HTTPHandler()) handlers.append(HTTPSHandler(ssl_factory)) if need_range_handler: handlers.extend( range_handlers ) handlers.append( auth_handler ) if self.opts.cache_openers: self._opener = CachedOpenerDirector(ssl_factory, *handlers) else: self._opener = ssl_factory.create_opener(*handlers) # OK, I don't like to do this, but otherwise, we end up with # TWO user-agent headers. self._opener.addheaders = [] return self._opener def _do_open(self): opener = self._get_opener() req = urllib2.Request(self.url, self.opts.data) # build request object self._add_headers(req) # add misc headers that we need self._build_range(req) # take care of reget and byterange stuff fo, hdr = self._make_request(req, opener) if self.reget_time and self.opts.reget == 'check_timestamp': # do this if we have a local file with known timestamp AND # we're in check_timestamp reget mode. fetch_again = 0 try: modified_tuple = hdr.getdate_tz('last-modified') modified_stamp = rfc822.mktime_tz(modified_tuple) if modified_stamp > self.reget_time: fetch_again = 1 except (TypeError,): fetch_again = 1 if fetch_again: # the server version is newer than the (incomplete) local # version, so we should abandon the version we're getting # and fetch the whole thing again. fo.close() self.opts.reget = None del req.headers['Range'] self._build_range(req) fo, hdr = self._make_request(req, opener) (scheme, host, path, parm, query, frag) = urlparse.urlparse(self.url) path = urllib.unquote(path) if not (self.opts.progress_obj or self.opts.raw_throttle() \ or self.opts.timeout): # if we're not using the progress_obj, throttling, or timeout # we can get a performance boost by going directly to # the underlying fileobject for reads. self.read = fo.read if hasattr(fo, 'readline'): self.readline = fo.readline elif self.opts.progress_obj: try: length = int(hdr['Content-Length']) length = length + self._amount_read # Account for regets except (KeyError, ValueError, TypeError): length = None self.opts.progress_obj.start(str(self.filename), urllib.unquote(self.url), os.path.basename(path), length, text=self.opts.text) self.opts.progress_obj.update(0) (self.fo, self.hdr) = (fo, hdr) def _add_headers(self, req): if self.opts.user_agent: req.add_header('User-agent', self.opts.user_agent) try: req_type = req.get_type() except ValueError: req_type = None if self.opts.http_headers and req_type in ('http', 'https'): for h, v in self.opts.http_headers: req.add_header(h, v) if self.opts.ftp_headers and req_type == 'ftp': for h, v in self.opts.ftp_headers: req.add_header(h, v) def _build_range(self, req): self.reget_time = None self.append = 0 reget_length = 0 rt = None if have_range and self.opts.reget and type(self.filename) == type(''): # we have reget turned on and we're dumping to a file try: s = os.stat(self.filename) except OSError: pass else: self.reget_time = s[ST_MTIME] reget_length = s[ST_SIZE] # Set initial length when regetting self._amount_read = reget_length rt = reget_length, '' self.append = 1 if self.opts.range: if not have_range: raise URLGrabError(10, _('Byte range requested but range '\ 'support unavailable')) rt = self.opts.range if rt[0]: rt = (rt[0] + reget_length, rt[1]) if rt: header = range_tuple_to_header(rt) if header: req.add_header('Range', header) def _make_request(self, req, opener): try: if have_socket_timeout and self.opts.timeout: old_to = socket.getdefaulttimeout() socket.setdefaulttimeout(self.opts.timeout) try: fo = opener.open(req) finally: socket.setdefaulttimeout(old_to) else: fo = opener.open(req) hdr = fo.info() except ValueError, e: raise URLGrabError(1, _('Bad URL: %s') % (e, )) except RangeError, e: raise URLGrabError(9, str(e)) except urllib2.HTTPError, e: new_e = URLGrabError(14, str(e)) new_e.code = e.code new_e.exception = e raise new_e except IOError, e: if hasattr(e, 'reason') and have_socket_timeout and \ isinstance(e.reason, TimeoutError): raise URLGrabError(12, _('Timeout: %s') % (e, )) else: raise URLGrabError(4, _('IOError: %s') % (e, )) except OSError, e: raise URLGrabError(5, _('OSError: %s') % (e, )) except HTTPException, e: raise URLGrabError(7, _('HTTP Exception (%s): %s') % \ (e.__class__.__name__, e)) else: return (fo, hdr) def _do_grab(self): """dump the file to self.filename.""" if self.append: new_fo = open(self.filename, 'ab') else: new_fo = open(self.filename, 'wb') bs = 1024*8 size = 0 block = self.read(bs) size = size + len(block) while block: new_fo.write(block) block = self.read(bs) size = size + len(block) new_fo.close() try: modified_tuple = self.hdr.getdate_tz('last-modified') modified_stamp = rfc822.mktime_tz(modified_tuple) os.utime(self.filename, (modified_stamp, modified_stamp)) except (TypeError,), e: pass return size def _fill_buffer(self, amt=None): """fill the buffer to contain at least 'amt' bytes by reading from the underlying file object. If amt is None, then it will read until it gets nothing more. It updates the progress meter and throttles after every self._rbufsize bytes.""" # the _rbuf test is only in this first 'if' for speed. It's not # logically necessary if self._rbuf and not amt is None: L = len(self._rbuf) if amt > L: amt = amt - L else: return # if we've made it here, then we don't have enough in the buffer # and we need to read more. buf = [self._rbuf] bufsize = len(self._rbuf) while amt is None or amt: # first, delay if necessary for throttling reasons if self.opts.raw_throttle(): diff = self._tsize/self.opts.raw_throttle() - \ (time.time() - self._ttime) if diff > 0: time.sleep(diff) self._ttime = time.time() # now read some data, up to self._rbufsize if amt is None: readamount = self._rbufsize else: readamount = min(amt, self._rbufsize) try: new = self.fo.read(readamount) except socket.error, e: raise URLGrabError(4, _('Socket Error: %s') % (e, )) except TimeoutError, e: raise URLGrabError(12, _('Timeout: %s') % (e, )) except IOError, e: raise URLGrabError(4, _('IOError: %s') %(e,)) newsize = len(new) if not newsize: break # no more to read if amt: amt = amt - newsize buf.append(new) bufsize = bufsize + newsize self._tsize = newsize self._amount_read = self._amount_read + newsize if self.opts.progress_obj: self.opts.progress_obj.update(self._amount_read) self._rbuf = string.join(buf, '') return def read(self, amt=None): self._fill_buffer(amt) if amt is None: s, self._rbuf = self._rbuf, '' else: s, self._rbuf = self._rbuf[:amt], self._rbuf[amt:] return s def readline(self, limit=-1): i = string.find(self._rbuf, '\n') while i < 0 and not (0 < limit <= len(self._rbuf)): L = len(self._rbuf) self._fill_buffer(L + self._rbufsize) if not len(self._rbuf) > L: break i = string.find(self._rbuf, '\n', L) if i < 0: i = len(self._rbuf) else: i = i+1 if 0 <= limit < len(self._rbuf): i = limit s, self._rbuf = self._rbuf[:i], self._rbuf[i:] return s def close(self): if self.opts.progress_obj: self.opts.progress_obj.end(self._amount_read) self.fo.close() if self.opts.close_connection: try: self.fo.close_connection() except: pass _handler_cache = [] def CachedOpenerDirector(ssl_factory = None, *handlers): for (cached_handlers, opener) in _handler_cache: if cached_handlers == handlers: for handler in opener.handlers: handler.add_parent(opener) return opener if not ssl_factory: ssl_factory = sslfactory.get_factory() opener = ssl_factory.create_opener(*handlers) _handler_cache.append( (handlers, opener) ) return opener _proxy_cache = [] def CachedProxyHandler(proxies): for (pdict, handler) in _proxy_cache: if pdict == proxies: if DEBUG: DEBUG.debug('re-using proxy settings: %s', proxies) break else: for k, v in proxies.items(): utype, url = urllib.splittype(v) host, other = urllib.splithost(url) if (utype is None) or (host is None): raise URLGrabError(13, _('Bad proxy URL: %s') % v) if DEBUG: DEBUG.info('creating new proxy handler: %s', proxies) handler = urllib2.ProxyHandler(proxies) _proxy_cache.append( (proxies, handler) ) return handler ##################################################################### # DEPRECATED FUNCTIONS def set_throttle(new_throttle): """Deprecated. Use: default_grabber.throttle = new_throttle""" default_grabber.throttle = new_throttle def set_bandwidth(new_bandwidth): """Deprecated. Use: default_grabber.bandwidth = new_bandwidth""" default_grabber.bandwidth = new_bandwidth def set_progress_obj(new_progress_obj): """Deprecated. Use: default_grabber.progress_obj = new_progress_obj""" default_grabber.progress_obj = new_progress_obj def set_user_agent(new_user_agent): """Deprecated. Use: default_grabber.user_agent = new_user_agent""" default_grabber.user_agent = new_user_agent def retrygrab(url, filename=None, copy_local=0, close_connection=0, progress_obj=None, throttle=None, bandwidth=None, numtries=3, retrycodes=[-1,2,4,5,6,7], checkfunc=None): """Deprecated. Use: urlgrab() with the retry arg instead""" kwargs = {'copy_local' : copy_local, 'close_connection' : close_connection, 'progress_obj' : progress_obj, 'throttle' : throttle, 'bandwidth' : bandwidth, 'retry' : numtries, 'retrycodes' : retrycodes, 'checkfunc' : checkfunc } return urlgrab(url, filename, **kwargs) ##################################################################### # TESTING def _main_test(): import sys try: url, filename = sys.argv[1:3] except ValueError: print 'usage:', sys.argv[0], \ '<url> <filename> [copy_local=0|1] [close_connection=0|1]' sys.exit() kwargs = {} for a in sys.argv[3:]: k, v = string.split(a, '=', 1) kwargs[k] = int(v) set_throttle(1.0) set_bandwidth(32 * 1024) print "throttle: %s, throttle bandwidth: %s B/s" % (default_grabber.throttle, default_grabber.bandwidth) try: from progress import text_progress_meter except ImportError, e: pass else: kwargs['progress_obj'] = text_progress_meter() try: name = apply(urlgrab, (url, filename), kwargs) except URLGrabError, e: print e else: print 'LOCAL FILE:', name def _retry_test(): import sys try: url, filename = sys.argv[1:3] except ValueError: print 'usage:', sys.argv[0], \ '<url> <filename> [copy_local=0|1] [close_connection=0|1]' sys.exit() kwargs = {} for a in sys.argv[3:]: k, v = string.split(a, '=', 1) kwargs[k] = int(v) try: from progress import text_progress_meter except ImportError, e: pass else: kwargs['progress_obj'] = text_progress_meter() def cfunc(filename, hello, there='foo'): print hello, there import random rnum = random.random() if rnum < .5: print 'forcing retry' raise URLGrabError(-1, 'forcing retry') if rnum < .75: print 'forcing failure' raise URLGrabError(-2, 'forcing immediate failure') print 'success' return kwargs['checkfunc'] = (cfunc, ('hello',), {'there':'there'}) try: name = apply(retrygrab, (url, filename), kwargs) except URLGrabError, e: print e else: print 'LOCAL FILE:', name def _file_object_test(filename=None): import random, cStringIO, sys if filename is None: filename = __file__ print 'using file "%s" for comparisons' % filename fo = open(filename) s_input = fo.read() fo.close() for testfunc in [_test_file_object_smallread, _test_file_object_readall, _test_file_object_readline, _test_file_object_readlines]: fo_input = cStringIO.StringIO(s_input) fo_output = cStringIO.StringIO() wrapper = URLGrabberFileObject(fo_input, None, 0) print 'testing %-30s ' % testfunc.__name__, testfunc(wrapper, fo_output) s_output = fo_output.getvalue() if s_output == s_input: print 'passed' else: print 'FAILED' def _test_file_object_smallread(wrapper, fo_output): while 1: s = wrapper.read(23) fo_output.write(s) if not s: return def _test_file_object_readall(wrapper, fo_output): s = wrapper.read() fo_output.write(s) def _test_file_object_readline(wrapper, fo_output): while 1: s = wrapper.readline() fo_output.write(s) if not s: return def _test_file_object_readlines(wrapper, fo_output): li = wrapper.readlines() fo_output.write(string.join(li, '')) if __name__ == '__main__': _main_test() _retry_test() _file_object_test('test') ########NEW FILE######## __FILENAME__ = keepalive # This library is free software; you can redistribute it and/or # modify it under the terms of the GNU Lesser General Public # License as published by the Free Software Foundation; either # version 2.1 of the License, or (at your option) any later version. # # This library is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # Lesser General Public License for more details. # # You should have received a copy of the GNU Lesser General Public # License along with this library; if not, write to the # Free Software Foundation, Inc., # 59 Temple Place, Suite 330, # Boston, MA 02111-1307 USA # This file is part of urlgrabber, a high-level cross-protocol url-grabber # Copyright 2002-2004 Michael D. Stenner, Ryan Tomayko """An HTTP handler for urllib2 that supports HTTP 1.1 and keepalive. >>> import urllib2 >>> from keepalive import HTTPHandler >>> keepalive_handler = HTTPHandler() >>> opener = urllib2.build_opener(keepalive_handler) >>> urllib2.install_opener(opener) >>> >>> fo = urllib2.urlopen('http://www.python.org') If a connection to a given host is requested, and all of the existing connections are still in use, another connection will be opened. If the handler tries to use an existing connection but it fails in some way, it will be closed and removed from the pool. To remove the handler, simply re-run build_opener with no arguments, and install that opener. You can explicitly close connections by using the close_connection() method of the returned file-like object (described below) or you can use the handler methods: close_connection(host) close_all() open_connections() NOTE: using the close_connection and close_all methods of the handler should be done with care when using multiple threads. * there is nothing that prevents another thread from creating new connections immediately after connections are closed * no checks are done to prevent in-use connections from being closed >>> keepalive_handler.close_all() EXTRA ATTRIBUTES AND METHODS Upon a status of 200, the object returned has a few additional attributes and methods, which should not be used if you want to remain consistent with the normal urllib2-returned objects: close_connection() - close the connection to the host readlines() - you know, readlines() status - the return status (ie 404) reason - english translation of status (ie 'File not found') If you want the best of both worlds, use this inside an AttributeError-catching try: >>> try: status = fo.status >>> except AttributeError: status = None Unfortunately, these are ONLY there if status == 200, so it's not easy to distinguish between non-200 responses. The reason is that urllib2 tries to do clever things with error codes 301, 302, 401, and 407, and it wraps the object upon return. For python versions earlier than 2.4, you can avoid this fancy error handling by setting the module-level global HANDLE_ERRORS to zero. You see, prior to 2.4, it's the HTTP Handler's job to determine what to handle specially, and what to just pass up. HANDLE_ERRORS == 0 means "pass everything up". In python 2.4, however, this job no longer belongs to the HTTP Handler and is now done by a NEW handler, HTTPErrorProcessor. Here's the bottom line: python version < 2.4 HANDLE_ERRORS == 1 (default) pass up 200, treat the rest as errors HANDLE_ERRORS == 0 pass everything up, error processing is left to the calling code python version >= 2.4 HANDLE_ERRORS == 1 pass up 200, treat the rest as errors HANDLE_ERRORS == 0 (default) pass everything up, let the other handlers (specifically, HTTPErrorProcessor) decide what to do In practice, setting the variable either way makes little difference in python 2.4, so for the most consistent behavior across versions, you probably just want to use the defaults, which will give you exceptions on errors. """ # $Id: keepalive.py,v 1.16 2006/09/22 00:58:05 mstenner Exp $ import urllib2 import httplib import socket import thread DEBUG = None import sslfactory import sys if sys.version_info < (2, 4): HANDLE_ERRORS = 1 else: HANDLE_ERRORS = 0 class ConnectionManager: """ The connection manager must be able to: * keep track of all existing """ def __init__(self): self._lock = thread.allocate_lock() self._hostmap = {} # map hosts to a list of connections self._connmap = {} # map connections to host self._readymap = {} # map connection to ready state def add(self, host, connection, ready): self._lock.acquire() try: if not self._hostmap.has_key(host): self._hostmap[host] = [] self._hostmap[host].append(connection) self._connmap[connection] = host self._readymap[connection] = ready finally: self._lock.release() def remove(self, connection): self._lock.acquire() try: try: host = self._connmap[connection] except KeyError: pass else: del self._connmap[connection] del self._readymap[connection] self._hostmap[host].remove(connection) if not self._hostmap[host]: del self._hostmap[host] finally: self._lock.release() def set_ready(self, connection, ready): try: self._readymap[connection] = ready except KeyError: pass def get_ready_conn(self, host): conn = None self._lock.acquire() try: if self._hostmap.has_key(host): for c in self._hostmap[host]: if self._readymap[c]: self._readymap[c] = 0 conn = c break finally: self._lock.release() return conn def get_all(self, host=None): if host: return list(self._hostmap.get(host, [])) else: return dict(self._hostmap) class KeepAliveHandler: def __init__(self): self._cm = ConnectionManager() #### Connection Management def open_connections(self): """return a list of connected hosts and the number of connections to each. [('foo.com:80', 2), ('bar.org', 1)]""" return [(host, len(li)) for (host, li) in self._cm.get_all().items()] def close_connection(self, host): """close connection(s) to <host> host is the host:port spec, as in 'www.cnn.com:8080' as passed in. no error occurs if there is no connection to that host.""" for h in self._cm.get_all(host): self._cm.remove(h) h.close() def close_all(self): """close all open connections""" for host, conns in self._cm.get_all().items(): for h in conns: self._cm.remove(h) h.close() def _request_closed(self, request, host, connection): """tells us that this request is now closed and the the connection is ready for another request""" self._cm.set_ready(connection, 1) def _remove_connection(self, host, connection, close=0): if close: connection.close() self._cm.remove(connection) #### Transaction Execution def do_open(self, req): host = req.get_host() if not host: raise urllib2.URLError('no host given') try: h = self._cm.get_ready_conn(host) while h: r = self._reuse_connection(h, req, host) # if this response is non-None, then it worked and we're # done. Break out, skipping the else block. if r: break # connection is bad - possibly closed by server # discard it and ask for the next free connection h.close() self._cm.remove(h) h = self._cm.get_ready_conn(host) else: # no (working) free connections were found. Create a new one. h = self._get_connection(host) if DEBUG: DEBUG.info("creating new connection to %s (%d)", host, id(h)) self._cm.add(host, h, 0) self._start_transaction(h, req) r = h.getresponse() except (socket.error, httplib.HTTPException), err: raise urllib2.URLError(err) # if not a persistent connection, don't try to reuse it if r.will_close: self._cm.remove(h) if DEBUG: DEBUG.info("STATUS: %s, %s", r.status, r.reason) r._handler = self r._host = host r._url = req.get_full_url() r._connection = h r.code = r.status r.headers = r.msg r.msg = r.reason if r.status == 200 or not HANDLE_ERRORS: return r else: return self.parent.error('http', req, r, r.status, r.msg, r.headers) def _reuse_connection(self, h, req, host): """start the transaction with a re-used connection return a response object (r) upon success or None on failure. This DOES not close or remove bad connections in cases where it returns. However, if an unexpected exception occurs, it will close and remove the connection before re-raising. """ try: self._start_transaction(h, req) r = h.getresponse() # note: just because we got something back doesn't mean it # worked. We'll check the version below, too. except (socket.error, httplib.HTTPException): r = None except: # adding this block just in case we've missed # something we will still raise the exception, but # lets try and close the connection and remove it # first. We previously got into a nasty loop # where an exception was uncaught, and so the # connection stayed open. On the next try, the # same exception was raised, etc. The tradeoff is # that it's now possible this call will raise # a DIFFERENT exception if DEBUG: DEBUG.error("unexpected exception - closing " + \ "connection to %s (%d)", host, id(h)) self._cm.remove(h) h.close() raise if r is None or r.version == 9: # httplib falls back to assuming HTTP 0.9 if it gets a # bad header back. This is most likely to happen if # the socket has been closed by the server since we # last used the connection. if DEBUG: DEBUG.info("failed to re-use connection to %s (%d)", host, id(h)) r = None else: if DEBUG: DEBUG.info("re-using connection to %s (%d)", host, id(h)) return r def _start_transaction(self, h, req): try: if req.has_data(): data = req.get_data() h.putrequest('POST', req.get_selector()) if not req.headers.has_key('Content-type'): h.putheader('Content-type', 'application/x-www-form-urlencoded') if not req.headers.has_key('Content-length'): h.putheader('Content-length', '%d' % len(data)) else: h.putrequest('GET', req.get_selector()) except (socket.error, httplib.HTTPException), err: raise urllib2.URLError(err) for args in self.parent.addheaders: h.putheader(*args) for k, v in req.headers.items(): h.putheader(k, v) h.endheaders() if req.has_data(): h.send(data) def _get_connection(self, host): return NotImplementedError class HTTPHandler(KeepAliveHandler, urllib2.HTTPHandler): def __init__(self): KeepAliveHandler.__init__(self) def http_open(self, req): return self.do_open(req) def _get_connection(self, host): return HTTPConnection(host) class HTTPSHandler(KeepAliveHandler, urllib2.HTTPSHandler): def __init__(self, ssl_factory=None): KeepAliveHandler.__init__(self) if not ssl_factory: ssl_factory = sslfactory.get_factory() self._ssl_factory = ssl_factory def https_open(self, req): return self.do_open(req) def _get_connection(self, host): return self._ssl_factory.get_https_connection(host) class HTTPResponse(httplib.HTTPResponse): # we need to subclass HTTPResponse in order to # 1) add readline() and readlines() methods # 2) add close_connection() methods # 3) add info() and geturl() methods # in order to add readline(), read must be modified to deal with a # buffer. example: readline must read a buffer and then spit back # one line at a time. The only real alternative is to read one # BYTE at a time (ick). Once something has been read, it can't be # put back (ok, maybe it can, but that's even uglier than this), # so if you THEN do a normal read, you must first take stuff from # the buffer. # the read method wraps the original to accomodate buffering, # although read() never adds to the buffer. # Both readline and readlines have been stolen with almost no # modification from socket.py def __init__(self, sock, debuglevel=0, strict=0, method=None): if method: # the httplib in python 2.3 uses the method arg httplib.HTTPResponse.__init__(self, sock, debuglevel, method) else: # 2.2 doesn't httplib.HTTPResponse.__init__(self, sock, debuglevel) self.fileno = sock.fileno self.code = None self._rbuf = '' self._rbufsize = 8096 self._handler = None # inserted by the handler later self._host = None # (same) self._url = None # (same) self._connection = None # (same) _raw_read = httplib.HTTPResponse.read def close(self): if self.fp: self.fp.close() self.fp = None if self._handler: self._handler._request_closed(self, self._host, self._connection) def close_connection(self): self._handler._remove_connection(self._host, self._connection, close=1) self.close() def info(self): return self.headers def geturl(self): return self._url def read(self, amt=None): # the _rbuf test is only in this first if for speed. It's not # logically necessary if self._rbuf and not amt is None: L = len(self._rbuf) if amt > L: amt -= L else: s = self._rbuf[:amt] self._rbuf = self._rbuf[amt:] return s s = self._rbuf + self._raw_read(amt) self._rbuf = '' return s def readline(self, limit=-1): data = "" i = self._rbuf.find('\n') while i < 0 and not (0 < limit <= len(self._rbuf)): new = self._raw_read(self._rbufsize) if not new: break i = new.find('\n') if i >= 0: i = i + len(self._rbuf) self._rbuf = self._rbuf + new if i < 0: i = len(self._rbuf) else: i = i+1 if 0 <= limit < len(self._rbuf): i = limit data, self._rbuf = self._rbuf[:i], self._rbuf[i:] return data def readlines(self, sizehint = 0): total = 0 list = [] while 1: line = self.readline() if not line: break list.append(line) total += len(line) if sizehint and total >= sizehint: break return list class HTTPConnection(httplib.HTTPConnection): # use the modified response class response_class = HTTPResponse class HTTPSConnection(httplib.HTTPSConnection): response_class = HTTPResponse ######################################################################### ##### TEST FUNCTIONS ######################################################################### def error_handler(url): global HANDLE_ERRORS orig = HANDLE_ERRORS keepalive_handler = HTTPHandler() opener = urllib2.build_opener(keepalive_handler) urllib2.install_opener(opener) pos = {0: 'off', 1: 'on'} for i in (0, 1): print " fancy error handling %s (HANDLE_ERRORS = %i)" % (pos[i], i) HANDLE_ERRORS = i try: fo = urllib2.urlopen(url) foo = fo.read() fo.close() try: status, reason = fo.status, fo.reason except AttributeError: status, reason = None, None except IOError, e: print " EXCEPTION: %s" % e raise else: print " status = %s, reason = %s" % (status, reason) HANDLE_ERRORS = orig hosts = keepalive_handler.open_connections() print "open connections:", hosts keepalive_handler.close_all() def continuity(url): import md5 format = '%25s: %s' # first fetch the file with the normal http handler opener = urllib2.build_opener() urllib2.install_opener(opener) fo = urllib2.urlopen(url) foo = fo.read() fo.close() m = md5.new(foo) print format % ('normal urllib', m.hexdigest()) # now install the keepalive handler and try again opener = urllib2.build_opener(HTTPHandler()) urllib2.install_opener(opener) fo = urllib2.urlopen(url) foo = fo.read() fo.close() m = md5.new(foo) print format % ('keepalive read', m.hexdigest()) fo = urllib2.urlopen(url) foo = '' while 1: f = fo.readline() if f: foo = foo + f else: break fo.close() m = md5.new(foo) print format % ('keepalive readline', m.hexdigest()) def comp(N, url): print ' making %i connections to:\n %s' % (N, url) sys.stdout.write(' first using the normal urllib handlers') # first use normal opener opener = urllib2.build_opener() urllib2.install_opener(opener) t1 = fetch(N, url) print ' TIME: %.3f s' % t1 sys.stdout.write(' now using the keepalive handler ') # now install the keepalive handler and try again opener = urllib2.build_opener(HTTPHandler()) urllib2.install_opener(opener) t2 = fetch(N, url) print ' TIME: %.3f s' % t2 print ' improvement factor: %.2f' % (t1/t2, ) def fetch(N, url, delay=0): import time lens = [] starttime = time.time() for i in range(N): if delay and i > 0: time.sleep(delay) fo = urllib2.urlopen(url) foo = fo.read() fo.close() lens.append(len(foo)) diff = time.time() - starttime j = 0 for i in lens[1:]: j = j + 1 if not i == lens[0]: print "WARNING: inconsistent length on read %i: %i" % (j, i) return diff def test_timeout(url): global DEBUG dbbackup = DEBUG class FakeLogger: def debug(self, msg, *args): print msg % args info = warning = error = debug DEBUG = FakeLogger() print " fetching the file to establish a connection" fo = urllib2.urlopen(url) data1 = fo.read() fo.close() i = 20 print " waiting %i seconds for the server to close the connection" % i while i > 0: sys.stdout.write('\r %2i' % i) sys.stdout.flush() time.sleep(1) i -= 1 sys.stderr.write('\r') print " fetching the file a second time" fo = urllib2.urlopen(url) data2 = fo.read() fo.close() if data1 == data2: print ' data are identical' else: print ' ERROR: DATA DIFFER' DEBUG = dbbackup def test(url, N=10): print "checking error hander (do this on a non-200)" try: error_handler(url) except IOError, e: print "exiting - exception will prevent further tests" sys.exit() print print "performing continuity test (making sure stuff isn't corrupted)" continuity(url) print print "performing speed comparison" comp(N, url) print print "performing dropped-connection check" test_timeout(url) if __name__ == '__main__': import time import sys try: N = int(sys.argv[1]) url = sys.argv[2] except: print "%s <integer> <url>" % sys.argv[0] else: test(url, N) ########NEW FILE######## __FILENAME__ = mirror # This library is free software; you can redistribute it and/or # modify it under the terms of the GNU Lesser General Public # License as published by the Free Software Foundation; either # version 2.1 of the License, or (at your option) any later version. # # This library is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # Lesser General Public License for more details. # # You should have received a copy of the GNU Lesser General Public # License along with this library; if not, write to the # Free Software Foundation, Inc., # 59 Temple Place, Suite 330, # Boston, MA 02111-1307 USA # This file is part of urlgrabber, a high-level cross-protocol url-grabber # Copyright 2002-2004 Michael D. Stenner, Ryan Tomayko """Module for downloading files from a pool of mirrors DESCRIPTION This module provides support for downloading files from a pool of mirrors with configurable failover policies. To a large extent, the failover policy is chosen by using different classes derived from the main class, MirrorGroup. Instances of MirrorGroup (and cousins) act very much like URLGrabber instances in that they have urlread, urlgrab, and urlopen methods. They can therefore, be used in very similar ways. from urlgrabber.grabber import URLGrabber from urlgrabber.mirror import MirrorGroup gr = URLGrabber() mg = MirrorGroup(gr, ['http://foo.com/some/directory/', 'http://bar.org/maybe/somewhere/else/', 'ftp://baz.net/some/other/place/entirely/'] mg.urlgrab('relative/path.zip') The assumption is that all mirrors are identical AFTER the base urls specified, so that any mirror can be used to fetch any file. FAILOVER The failover mechanism is designed to be customized by subclassing from MirrorGroup to change the details of the behavior. In general, the classes maintain a master mirror list and a "current mirror" index. When a download is initiated, a copy of this list and index is created for that download only. The specific failover policy depends on the class used, and so is documented in the class documentation. Note that ANY behavior of the class can be overridden, so any failover policy at all is possible (although you may need to change the interface in extreme cases). CUSTOMIZATION Most customization of a MirrorGroup object is done at instantiation time (or via subclassing). There are four major types of customization: 1) Pass in a custom urlgrabber - The passed in urlgrabber will be used (by default... see #2) for the grabs, so options to it apply for the url-fetching 2) Custom mirror list - Mirror lists can simply be a list of stings mirrors (as shown in the example above) but each can also be a dict, allowing for more options. For example, the first mirror in the list above could also have been: {'mirror': 'http://foo.com/some/directory/', 'grabber': <a custom grabber to be used for this mirror>, 'kwargs': { <a dict of arguments passed to the grabber> }} All mirrors are converted to this format internally. If 'grabber' is omitted, the default grabber will be used. If kwargs are omitted, then (duh) they will not be used. 3) Pass keyword arguments when instantiating the mirror group. See, for example, the failure_callback argument. 4) Finally, any kwargs passed in for the specific file (to the urlgrab method, for example) will be folded in. The options passed into the grabber's urlXXX methods will override any options specified in a custom mirror dict. """ # $Id: mirror.py,v 1.14 2006/02/22 18:26:46 mstenner Exp $ import random import thread # needed for locking to make this threadsafe from grabber import URLGrabError, CallbackObject, DEBUG try: from i18n import _ except ImportError, msg: def _(st): return st class GrabRequest: """This is a dummy class used to hold information about the specific request. For example, a single file. By maintaining this information separately, we can accomplish two things: 1) make it a little easier to be threadsafe 2) have request-specific parameters """ pass class MirrorGroup: """Base Mirror class Instances of this class are built with a grabber object and a list of mirrors. Then all calls to urlXXX should be passed relative urls. The requested file will be searched for on the first mirror. If the grabber raises an exception (possibly after some retries) then that mirror will be removed from the list, and the next will be attempted. If all mirrors are exhausted, then an exception will be raised. MirrorGroup has the following failover policy: * downloads begin with the first mirror * by default (see default_action below) a failure (after retries) causes it to increment the local AND master indices. Also, the current mirror is removed from the local list (but NOT the master list - the mirror can potentially be used for other files) * if the local list is ever exhausted, a URLGrabError will be raised (errno=256, no more mirrors) OPTIONS In addition to the required arguments "grabber" and "mirrors", MirrorGroup also takes the following optional arguments: default_action A dict that describes the actions to be taken upon failure (after retries). default_action can contain any of the following keys (shown here with their default values): default_action = {'increment': 1, 'increment_master': 1, 'remove': 1, 'remove_master': 0, 'fail': 0} In this context, 'increment' means "use the next mirror" and 'remove' means "never use this mirror again". The two 'master' values refer to the instance-level mirror list (used for all files), whereas the non-master values refer to the current download only. The 'fail' option will cause immediate failure by re-raising the exception and no further attempts to get the current download. This dict can be set at instantiation time, mg = MirrorGroup(grabber, mirrors, default_action={'fail':1}) at method-execution time (only applies to current fetch), filename = mg.urlgrab(url, default_action={'increment': 0}) or by returning an action dict from the failure_callback return {'fail':0} in increasing precedence. If all three of these were done, the net result would be: {'increment': 0, # set in method 'increment_master': 1, # class default 'remove': 1, # class default 'remove_master': 0, # class default 'fail': 0} # set at instantiation, reset # from callback failure_callback this is a callback that will be called when a mirror "fails", meaning the grabber raises some URLGrabError. If this is a tuple, it is interpreted to be of the form (cb, args, kwargs) where cb is the actual callable object (function, method, etc). Otherwise, it is assumed to be the callable object itself. The callback will be passed a grabber.CallbackObject instance along with args and kwargs (if present). The following attributes are defined withing the instance: obj.exception = < exception that was raised > obj.mirror = < the mirror that was tried > obj.relative_url = < url relative to the mirror > obj.url = < full url that failed > # .url is just the combination of .mirror # and .relative_url The failure callback can return an action dict, as described above. Like default_action, the failure_callback can be set at instantiation time or when the urlXXX method is called. In the latter case, it applies only for that fetch. The callback can re-raise the exception quite easily. For example, this is a perfectly adequate callback function: def callback(obj): raise obj.exception WARNING: do not save the exception object (or the CallbackObject instance). As they contain stack frame references, they can lead to circular references. Notes: * The behavior can be customized by deriving and overriding the 'CONFIGURATION METHODS' * The 'grabber' instance is kept as a reference, not copied. Therefore, the grabber instance can be modified externally and changes will take effect immediately. """ # notes on thread-safety: # A GrabRequest should never be shared by multiple threads because # it's never saved inside the MG object and never returned outside it. # therefore, it should be safe to access/modify grabrequest data # without a lock. However, accessing the mirrors and _next attributes # of the MG itself must be done when locked to prevent (for example) # removal of the wrong mirror. ############################################################## # CONFIGURATION METHODS - intended to be overridden to # customize behavior def __init__(self, grabber, mirrors, **kwargs): """Initialize the MirrorGroup object. REQUIRED ARGUMENTS grabber - URLGrabber instance mirrors - a list of mirrors OPTIONAL ARGUMENTS failure_callback - callback to be used when a mirror fails default_action - dict of failure actions See the module-level and class level documentation for more details. """ # OVERRIDE IDEAS: # shuffle the list to randomize order self.grabber = grabber self.mirrors = self._parse_mirrors(mirrors) self._next = 0 self._lock = thread.allocate_lock() self.default_action = None self._process_kwargs(kwargs) # if these values are found in **kwargs passed to one of the urlXXX # methods, they will be stripped before getting passed on to the # grabber options = ['default_action', 'failure_callback'] def _process_kwargs(self, kwargs): self.failure_callback = kwargs.get('failure_callback') self.default_action = kwargs.get('default_action') def _parse_mirrors(self, mirrors): parsed_mirrors = [] for m in mirrors: if type(m) == type(''): m = {'mirror': m} parsed_mirrors.append(m) return parsed_mirrors def _load_gr(self, gr): # OVERRIDE IDEAS: # shuffle gr list self._lock.acquire() gr.mirrors = list(self.mirrors) gr._next = self._next self._lock.release() def _get_mirror(self, gr): # OVERRIDE IDEAS: # return a random mirror so that multiple mirrors get used # even without failures. if not gr.mirrors: raise URLGrabError(256, _('No more mirrors to try.')) return gr.mirrors[gr._next] def _failure(self, gr, cb_obj): # OVERRIDE IDEAS: # inspect the error - remove=1 for 404, remove=2 for connection # refused, etc. (this can also be done via # the callback) cb = gr.kw.get('failure_callback') or self.failure_callback if cb: if type(cb) == type( () ): cb, args, kwargs = cb else: args, kwargs = (), {} action = cb(cb_obj, *args, **kwargs) or {} else: action = {} # XXXX - decide - there are two ways to do this # the first is action-overriding as a whole - use the entire action # or fall back on module level defaults #action = action or gr.kw.get('default_action') or self.default_action # the other is to fall through for each element in the action dict a = dict(self.default_action or {}) a.update(gr.kw.get('default_action', {})) a.update(action) action = a self.increment_mirror(gr, action) if action and action.get('fail', 0): raise def increment_mirror(self, gr, action={}): """Tell the mirror object increment the mirror index This increments the mirror index, which amounts to telling the mirror object to use a different mirror (for this and future downloads). This is a SEMI-public method. It will be called internally, and you may never need to call it. However, it is provided (and is made public) so that the calling program can increment the mirror choice for methods like urlopen. For example, with urlopen, there's no good way for the mirror group to know that an error occurs mid-download (it's already returned and given you the file object). remove --- can have several values 0 do not remove the mirror from the list 1 remove the mirror for this download only 2 remove the mirror permanently beware of remove=0 as it can lead to infinite loops """ badmirror = gr.mirrors[gr._next] self._lock.acquire() try: ind = self.mirrors.index(badmirror) except ValueError: pass else: if action.get('remove_master', 0): del self.mirrors[ind] elif self._next == ind and action.get('increment_master', 1): self._next += 1 if self._next >= len(self.mirrors): self._next = 0 self._lock.release() if action.get('remove', 1): del gr.mirrors[gr._next] elif action.get('increment', 1): gr._next += 1 if gr._next >= len(gr.mirrors): gr._next = 0 if DEBUG: grm = [m['mirror'] for m in gr.mirrors] DEBUG.info('GR mirrors: [%s] %i', ' '.join(grm), gr._next) selfm = [m['mirror'] for m in self.mirrors] DEBUG.info('MAIN mirrors: [%s] %i', ' '.join(selfm), self._next) ##################################################################### # NON-CONFIGURATION METHODS # these methods are designed to be largely workhorse methods that # are not intended to be overridden. That doesn't mean you can't; # if you want to, feel free, but most things can be done by # by overriding the configuration methods :) def _join_url(self, base_url, rel_url): if base_url.endswith('/') or rel_url.startswith('/'): return base_url + rel_url else: return base_url + '/' + rel_url def _mirror_try(self, func, url, kw): gr = GrabRequest() gr.func = func gr.url = url gr.kw = dict(kw) self._load_gr(gr) for k in self.options: try: del kw[k] except KeyError: pass while 1: mirrorchoice = self._get_mirror(gr) fullurl = self._join_url(mirrorchoice['mirror'], gr.url) kwargs = dict(mirrorchoice.get('kwargs', {})) kwargs.update(kw) grabber = mirrorchoice.get('grabber') or self.grabber func_ref = getattr(grabber, func) if DEBUG: DEBUG.info('MIRROR: trying %s -> %s', url, fullurl) try: return func_ref( *(fullurl,), **kwargs ) except URLGrabError, e: if DEBUG: DEBUG.info('MIRROR: failed') obj = CallbackObject() obj.exception = e obj.mirror = mirrorchoice['mirror'] obj.relative_url = gr.url obj.url = fullurl self._failure(gr, obj) def urlgrab(self, url, filename=None, **kwargs): kw = dict(kwargs) kw['filename'] = filename func = 'urlgrab' return self._mirror_try(func, url, kw) def urlopen(self, url, **kwargs): kw = dict(kwargs) func = 'urlopen' return self._mirror_try(func, url, kw) def urlread(self, url, limit=None, **kwargs): kw = dict(kwargs) kw['limit'] = limit func = 'urlread' return self._mirror_try(func, url, kw) class MGRandomStart(MirrorGroup): """A mirror group that starts at a random mirror in the list. This behavior of this class is identical to MirrorGroup, except that it starts at a random location in the mirror list. """ def __init__(self, grabber, mirrors, **kwargs): """Initialize the object The arguments for intialization are the same as for MirrorGroup """ MirrorGroup.__init__(self, grabber, mirrors, **kwargs) self._next = random.randrange(len(mirrors)) class MGRandomOrder(MirrorGroup): """A mirror group that uses mirrors in a random order. This behavior of this class is identical to MirrorGroup, except that it uses the mirrors in a random order. Note that the order is set at initialization time and fixed thereafter. That is, it does not pick a random mirror after each failure. """ def __init__(self, grabber, mirrors, **kwargs): """Initialize the object The arguments for intialization are the same as for MirrorGroup """ MirrorGroup.__init__(self, grabber, mirrors, **kwargs) random.shuffle(self.mirrors) if __name__ == '__main__': pass ########NEW FILE######## __FILENAME__ = progress # This library is free software; you can redistribute it and/or # modify it under the terms of the GNU Lesser General Public # License as published by the Free Software Foundation; either # version 2.1 of the License, or (at your option) any later version. # # This library is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # Lesser General Public License for more details. # # You should have received a copy of the GNU Lesser General Public # License along with this library; if not, write to the # Free Software Foundation, Inc., # 59 Temple Place, Suite 330, # Boston, MA 02111-1307 USA # This file is part of urlgrabber, a high-level cross-protocol url-grabber # Copyright 2002-2004 Michael D. Stenner, Ryan Tomayko # $Id: progress.py,v 1.7 2005/08/19 21:59:07 mstenner Exp $ import sys import time import math import thread class BaseMeter: def __init__(self): self.update_period = 0.3 # seconds self.filename = None self.url = None self.basename = None self.text = None self.size = None self.start_time = None self.last_amount_read = 0 self.last_update_time = None self.re = RateEstimator() def start(self, filename=None, url=None, basename=None, size=None, now=None, text=None): self.filename = filename self.url = url self.basename = basename self.text = text #size = None ######### TESTING self.size = size if not size is None: self.fsize = format_number(size) + 'B' if now is None: now = time.time() self.start_time = now self.re.start(size, now) self.last_amount_read = 0 self.last_update_time = now self._do_start(now) def _do_start(self, now=None): pass def update(self, amount_read, now=None): # for a real gui, you probably want to override and put a call # to your mainloop iteration function here if now is None: now = time.time() if (now >= self.last_update_time + self.update_period) or \ not self.last_update_time: self.re.update(amount_read, now) self.last_amount_read = amount_read self.last_update_time = now self._do_update(amount_read, now) def _do_update(self, amount_read, now=None): pass def end(self, amount_read, now=None): if now is None: now = time.time() self.re.update(amount_read, now) self.last_amount_read = amount_read self.last_update_time = now self._do_end(amount_read, now) def _do_end(self, amount_read, now=None): pass class TextMeter(BaseMeter): def __init__(self, fo=sys.stderr): BaseMeter.__init__(self) self.fo = fo def _do_update(self, amount_read, now=None): etime = self.re.elapsed_time() fetime = format_time(etime) fread = format_number(amount_read) #self.size = None if self.text is not None: text = self.text else: text = self.basename if self.size is None: out = '\r%-60.60s %5sB %s ' % \ (text, fread, fetime) else: rtime = self.re.remaining_time() frtime = format_time(rtime) frac = self.re.fraction_read() bar = '='*int(25 * frac) out = '\r%-25.25s %3i%% |%-25.25s| %5sB %8s ETA ' % \ (text, frac*100, bar, fread, frtime) self.fo.write(out) self.fo.flush() def _do_end(self, amount_read, now=None): total_time = format_time(self.re.elapsed_time()) total_size = format_number(amount_read) if self.text is not None: text = self.text else: text = self.basename if self.size is None: out = '\r%-60.60s %5sB %s ' % \ (text, total_size, total_time) else: bar = '='*25 out = '\r%-25.25s %3i%% |%-25.25s| %5sB %8s ' % \ (text, 100, bar, total_size, total_time) self.fo.write(out + '\n') self.fo.flush() text_progress_meter = TextMeter class MultiFileHelper(BaseMeter): def __init__(self, master): BaseMeter.__init__(self) self.master = master def _do_start(self, now): self.master.start_meter(self, now) def _do_update(self, amount_read, now): # elapsed time since last update self.master.update_meter(self, now) def _do_end(self, amount_read, now): self.ftotal_time = format_time(now - self.start_time) self.ftotal_size = format_number(self.last_amount_read) self.master.end_meter(self, now) def failure(self, message, now=None): self.master.failure_meter(self, message, now) def message(self, message): self.master.message_meter(self, message) class MultiFileMeter: helperclass = MultiFileHelper def __init__(self): self.meters = [] self.in_progress_meters = [] self._lock = thread.allocate_lock() self.update_period = 0.3 # seconds self.numfiles = None self.finished_files = 0 self.failed_files = 0 self.open_files = 0 self.total_size = None self.failed_size = 0 self.start_time = None self.finished_file_size = 0 self.last_update_time = None self.re = RateEstimator() def start(self, numfiles=None, total_size=None, now=None): if now is None: now = time.time() self.numfiles = numfiles self.finished_files = 0 self.failed_files = 0 self.open_files = 0 self.total_size = total_size self.failed_size = 0 self.start_time = now self.finished_file_size = 0 self.last_update_time = now self.re.start(total_size, now) self._do_start(now) def _do_start(self, now): pass def end(self, now=None): if now is None: now = time.time() self._do_end(now) def _do_end(self, now): pass def lock(self): self._lock.acquire() def unlock(self): self._lock.release() ########################################################### # child meter creation and destruction def newMeter(self): newmeter = self.helperclass(self) self.meters.append(newmeter) return newmeter def removeMeter(self, meter): self.meters.remove(meter) ########################################################### # child functions - these should only be called by helpers def start_meter(self, meter, now): if not meter in self.meters: raise ValueError('attempt to use orphaned meter') self._lock.acquire() try: if not meter in self.in_progress_meters: self.in_progress_meters.append(meter) self.open_files += 1 finally: self._lock.release() self._do_start_meter(meter, now) def _do_start_meter(self, meter, now): pass def update_meter(self, meter, now): if not meter in self.meters: raise ValueError('attempt to use orphaned meter') if (now >= self.last_update_time + self.update_period) or \ not self.last_update_time: self.re.update(self._amount_read(), now) self.last_update_time = now self._do_update_meter(meter, now) def _do_update_meter(self, meter, now): pass def end_meter(self, meter, now): if not meter in self.meters: raise ValueError('attempt to use orphaned meter') self._lock.acquire() try: try: self.in_progress_meters.remove(meter) except ValueError: pass self.open_files -= 1 self.finished_files += 1 self.finished_file_size += meter.last_amount_read finally: self._lock.release() self._do_end_meter(meter, now) def _do_end_meter(self, meter, now): pass def failure_meter(self, meter, message, now): if not meter in self.meters: raise ValueError('attempt to use orphaned meter') self._lock.acquire() try: try: self.in_progress_meters.remove(meter) except ValueError: pass self.open_files -= 1 self.failed_files += 1 if meter.size and self.failed_size is not None: self.failed_size += meter.size else: self.failed_size = None finally: self._lock.release() self._do_failure_meter(meter, message, now) def _do_failure_meter(self, meter, message, now): pass def message_meter(self, meter, message): pass ######################################################## # internal functions def _amount_read(self): tot = self.finished_file_size for m in self.in_progress_meters: tot += m.last_amount_read return tot class TextMultiFileMeter(MultiFileMeter): def __init__(self, fo=sys.stderr): self.fo = fo MultiFileMeter.__init__(self) # files: ###/### ###% data: ######/###### ###% time: ##:##:##/##:##:## def _do_update_meter(self, meter, now): self._lock.acquire() try: format = "files: %3i/%-3i %3i%% data: %6.6s/%-6.6s %3i%% " \ "time: %8.8s/%8.8s" df = self.finished_files tf = self.numfiles or 1 pf = 100 * float(df)/tf + 0.49 dd = self.re.last_amount_read td = self.total_size pd = 100 * (self.re.fraction_read() or 0) + 0.49 dt = self.re.elapsed_time() rt = self.re.remaining_time() if rt is None: tt = None else: tt = dt + rt fdd = format_number(dd) + 'B' ftd = format_number(td) + 'B' fdt = format_time(dt, 1) ftt = format_time(tt, 1) out = '%-79.79s' % (format % (df, tf, pf, fdd, ftd, pd, fdt, ftt)) self.fo.write('\r' + out) self.fo.flush() finally: self._lock.release() def _do_end_meter(self, meter, now): self._lock.acquire() try: format = "%-30.30s %6.6s %8.8s %9.9s" fn = meter.basename size = meter.last_amount_read fsize = format_number(size) + 'B' et = meter.re.elapsed_time() fet = format_time(et, 1) frate = format_number(size / et) + 'B/s' out = '%-79.79s' % (format % (fn, fsize, fet, frate)) self.fo.write('\r' + out + '\n') finally: self._lock.release() self._do_update_meter(meter, now) def _do_failure_meter(self, meter, message, now): self._lock.acquire() try: format = "%-30.30s %6.6s %s" fn = meter.basename if type(message) in (type(''), type(u'')): message = message.splitlines() if not message: message = [''] out = '%-79s' % (format % (fn, 'FAILED', message[0] or '')) self.fo.write('\r' + out + '\n') for m in message[1:]: self.fo.write(' ' + m + '\n') self._lock.release() finally: self._do_update_meter(meter, now) def message_meter(self, meter, message): self._lock.acquire() try: pass finally: self._lock.release() def _do_end(self, now): self._do_update_meter(None, now) self._lock.acquire() try: self.fo.write('\n') self.fo.flush() finally: self._lock.release() ###################################################################### # support classes and functions class RateEstimator: def __init__(self, timescale=5.0): self.timescale = timescale def start(self, total=None, now=None): if now is None: now = time.time() self.total = total self.start_time = now self.last_update_time = now self.last_amount_read = 0 self.ave_rate = None def update(self, amount_read, now=None): if now is None: now = time.time() if amount_read == 0: # if we just started this file, all bets are off self.last_update_time = now self.last_amount_read = 0 self.ave_rate = None return #print 'times', now, self.last_update_time time_diff = now - self.last_update_time read_diff = amount_read - self.last_amount_read self.last_update_time = now self.last_amount_read = amount_read self.ave_rate = self._temporal_rolling_ave(\ time_diff, read_diff, self.ave_rate, self.timescale) #print 'results', time_diff, read_diff, self.ave_rate ##################################################################### # result methods def average_rate(self): "get the average transfer rate (in bytes/second)" return self.ave_rate def elapsed_time(self): "the time between the start of the transfer and the most recent update" return self.last_update_time - self.start_time def remaining_time(self): "estimated time remaining" if not self.ave_rate or not self.total: return None return (self.total - self.last_amount_read) / self.ave_rate def fraction_read(self): """the fraction of the data that has been read (can be None for unknown transfer size)""" if self.total is None: return None elif self.total == 0: return 1.0 else: return float(self.last_amount_read)/self.total ######################################################################### # support methods def _temporal_rolling_ave(self, time_diff, read_diff, last_ave, timescale): """a temporal rolling average performs smooth averaging even when updates come at irregular intervals. This is performed by scaling the "epsilon" according to the time since the last update. Specifically, epsilon = time_diff / timescale As a general rule, the average will take on a completely new value after 'timescale' seconds.""" epsilon = time_diff / timescale if epsilon > 1: epsilon = 1.0 return self._rolling_ave(time_diff, read_diff, last_ave, epsilon) def _rolling_ave(self, time_diff, read_diff, last_ave, epsilon): """perform a "rolling average" iteration a rolling average "folds" new data into an existing average with some weight, epsilon. epsilon must be between 0.0 and 1.0 (inclusive) a value of 0.0 means only the old value (initial value) counts, and a value of 1.0 means only the newest value is considered.""" try: recent_rate = read_diff / time_diff except ZeroDivisionError: recent_rate = None if last_ave is None: return recent_rate elif recent_rate is None: return last_ave # at this point, both last_ave and recent_rate are numbers return epsilon * recent_rate + (1 - epsilon) * last_ave def _round_remaining_time(self, rt, start_time=15.0): """round the remaining time, depending on its size If rt is between n*start_time and (n+1)*start_time round downward to the nearest multiple of n (for any counting number n). If rt < start_time, round down to the nearest 1. For example (for start_time = 15.0): 2.7 -> 2.0 25.2 -> 25.0 26.4 -> 26.0 35.3 -> 34.0 63.6 -> 60.0 """ if rt < 0: return 0.0 shift = int(math.log(rt/start_time)/math.log(2)) rt = int(rt) if shift <= 0: return rt return float(int(rt) >> shift << shift) def format_time(seconds, use_hours=0): if seconds is None or seconds < 0: if use_hours: return '--:--:--' else: return '--:--' else: seconds = int(seconds) minutes = seconds / 60 seconds = seconds % 60 if use_hours: hours = minutes / 60 minutes = minutes % 60 return '%02i:%02i:%02i' % (hours, minutes, seconds) else: return '%02i:%02i' % (minutes, seconds) def format_number(number, SI=0, space=' '): """Turn numbers into human-readable metric-like numbers""" symbols = ['', # (none) 'k', # kilo 'M', # mega 'G', # giga 'T', # tera 'P', # peta 'E', # exa 'Z', # zetta 'Y'] # yotta if SI: step = 1000.0 else: step = 1024.0 thresh = 999 depth = 0 max_depth = len(symbols) - 1 # we want numbers between 0 and thresh, but don't exceed the length # of our list. In that event, the formatting will be screwed up, # but it'll still show the right number. while number > thresh and depth < max_depth: depth = depth + 1 number = number / step if type(number) == type(1) or type(number) == type(1L): # it's an int or a long, which means it didn't get divided, # which means it's already short enough format = '%i%s%s' elif number < 9.95: # must use 9.95 for proper sizing. For example, 9.99 will be # rounded to 10.0 with the .1f format string (which is too long) format = '%.1f%s%s' else: format = '%.0f%s%s' return(format % (float(number or 0), space, symbols[depth])) ########NEW FILE######## __FILENAME__ = sslfactory # This library is free software; you can redistribute it and/or # modify it under the terms of the GNU Lesser General Public # License as published by the Free Software Foundation; either # version 2.1 of the License, or (at your option) any later version. # # This library is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # Lesser General Public License for more details. # # You should have received a copy of the GNU Lesser General Public # License along with this library; if not, write to the # Free Software Foundation, Inc., # 59 Temple Place, Suite 330, # Boston, MA 02111-1307 USA # This file is part of urlgrabber, a high-level cross-protocol url-grabber import httplib import urllib2 try: from M2Crypto import SSL from M2Crypto import httpslib from M2Crypto import m2urllib2 SSL.Connection.clientPostConnectionCheck = None have_m2crypto = True except ImportError: have_m2crypto = False DEBUG = None if have_m2crypto: class M2SSLFactory: def __init__(self, ssl_ca_cert, ssl_context): self.ssl_context = self._get_ssl_context(ssl_ca_cert, ssl_context) def _get_ssl_context(self, ssl_ca_cert, ssl_context): """ Create an ssl context using the CA cert file or ssl context. The CA cert is used first if it was passed as an option. If not, then the supplied ssl context is used. If no ssl context was supplied, None is returned. """ if ssl_ca_cert: context = SSL.Context() context.load_verify_locations(ssl_ca_cert) context.set_verify(SSL.verify_none, -1) return context else: return ssl_context def create_https_connection(self, host, response_class = None): connection = httplib.HTTPSConnection(host, self.ssl_context) if response_class: connection.response_class = response_class return connection def create_opener(self, *handlers): return m2urllib2.build_opener(self.ssl_context, *handlers) class SSLFactory: def create_https_connection(self, host, response_class = None): connection = httplib.HTTPSConnection(host) if response_class: connection.response_class = response_class return connection def create_opener(self, *handlers): return urllib2.build_opener(*handlers) def get_factory(ssl_ca_cert = None, ssl_context = None): """ Return an SSLFactory, based on if M2Crypto is available. """ if have_m2crypto: return M2SSLFactory(ssl_ca_cert, ssl_context) else: # Log here if someone provides the args but we don't use them. if ssl_ca_cert or ssl_context: if DEBUG: DEBUG.warning("SSL arguments supplied, but M2Crypto is not available. " "Using Python SSL.") return SSLFactory() ########NEW FILE######## __FILENAME__ = version # # Chris Lumens <clumens@redhat.com> # # Copyright 2006, 2007, 2008, 2009, 2010 Red Hat, Inc. # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. Any Red Hat # trademarks that are incorporated in the source code or documentation are not # subject to the GNU General Public License and may only be used or replicated # with the express permission of Red Hat, Inc. # """ Methods for working with kickstart versions. This module defines several symbolic constants that specify kickstart syntax versions. Each version corresponds roughly to one release of Red Hat Linux, Red Hat Enterprise Linux, or Fedora Core as these are where most syntax changes take place. This module also exports several functions: makeVersion - Given a version number, return an instance of the matching handler class. returnClassForVersion - Given a version number, return the matching handler class. This does not return an instance of that class, however. stringToVersion - Convert a string representation of a version number into the symbolic constant. versionToString - Perform the reverse mapping. versionFromFile - Read a kickstart file and determine the version of syntax it uses. This requires the kickstart file to have a version= comment in it. """ import imputil, re, sys from urlgrabber import urlopen import gettext _ = lambda x: gettext.ldgettext("pykickstart", x) from pykickstart.errors import KickstartVersionError # Symbolic names for internal version numbers. RHEL3 = 900 FC3 = 1000 RHEL4 = 1100 FC4 = 2000 FC5 = 3000 FC6 = 4000 RHEL5 = 4100 F7 = 5000 F8 = 6000 F9 = 7000 F10 = 8000 F11 = 9000 F12 = 10000 F13 = 11000 RHEL6 = 11100 F14 = 12000 F15 = 13000 F16 = 14000 # This always points at the latest version and is the default. DEVEL = F16 # A one-to-one mapping from string representations to version numbers. versionMap = { "DEVEL": DEVEL, "FC3": FC3, "FC4": FC4, "FC5": FC5, "FC6": FC6, "F7": F7, "F8": F8, "F9": F9, "F10": F10, "F11": F11, "F12": F12, "F13": F13, "F14": F14, "F15": F15, "F16": F16, "RHEL3": RHEL3, "RHEL4": RHEL4, "RHEL5": RHEL5, "RHEL6": RHEL6 } def stringToVersion(s): """Convert string into one of the provided version constants. Raises KickstartVersionError if string does not match anything. """ # First try these short forms. try: return versionMap[s.upper()] except KeyError: pass # Now try the Fedora versions. m = re.match("^fedora.* (\d+)$", s, re.I) if m and m.group(1): if versionMap.has_key("FC" + m.group(1)): return versionMap["FC" + m.group(1)] elif versionMap.has_key("F" + m.group(1)): return versionMap["F" + m.group(1)] else: raise KickstartVersionError(_("Unsupported version specified: %s") % s) # Now try the RHEL versions. m = re.match("^red hat enterprise linux.* (\d+)([\.\d]*)$", s, re.I) if m and m.group(1): if versionMap.has_key("RHEL" + m.group(1)): return versionMap["RHEL" + m.group(1)] else: raise KickstartVersionError(_("Unsupported version specified: %s") % s) # If nothing else worked, we're out of options. raise KickstartVersionError(_("Unsupported version specified: %s") % s) def versionToString(version, skipDevel=False): """Convert version into a string representation of the version number. This is the reverse operation of stringToVersion. Raises KickstartVersionError if version does not match anything. """ if not skipDevel and version == versionMap["DEVEL"]: return "DEVEL" for (key, val) in versionMap.iteritems(): if key == "DEVEL": continue elif val == version: return key raise KickstartVersionError(_("Unsupported version specified: %s") % version) def versionFromFile(f): """Given a file or URL, look for a line starting with #version= and return the version number. If no version is found, return DEVEL. """ v = DEVEL fh = urlopen(f) while True: try: l = fh.readline() except StopIteration: break # At the end of the file? if l == "": break if l.isspace() or l.strip() == "": continue if l[:9] == "#version=": v = stringToVersion(l[9:].rstrip()) break fh.close() return v def returnClassForVersion(version=DEVEL): """Return the class of the syntax handler for version. version can be either a string or the matching constant. Raises KickstartValueError if version does not match anything. """ try: version = int(version) module = "%s" % versionToString(version, skipDevel=True) except ValueError: module = "%s" % version version = stringToVersion(version) module = module.lower() try: import pykickstart.handlers sys.path.extend(pykickstart.handlers.__path__) found = imputil.imp.find_module(module) loaded = imputil.imp.load_module(module, found[0], found[1], found[2]) for (k, v) in loaded.__dict__.iteritems(): if k.lower().endswith("%shandler" % module): return v except: raise KickstartVersionError(_("Unsupported version specified: %s") % version) def makeVersion(version=DEVEL): """Return a new instance of the syntax handler for version. version can be either a string or the matching constant. This function is useful for standalone programs which just need to handle a specific version of kickstart syntax (as provided by a command line argument, for example) and need to instantiate the correct object. """ cl = returnClassForVersion(version) return cl() ########NEW FILE######## __FILENAME__ = bootstrap #!/usr/bin/python -tt # # Copyright (c) 2009, 2010, 2011 Intel, Inc. # # This program is free software; you can redistribute it and/or modify it # under the terms of the GNU General Public License as published by the Free # Software Foundation; version 2 of the License # # This program is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY # or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License # for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., 59 # Temple Place - Suite 330, Boston, MA 02111-1307, USA. from __future__ import with_statement import os import sys import tempfile import shutil import subprocess import rpm from mic import msger from mic.utils import errors, proxy, misc from mic.utils.rpmmisc import readRpmHeader, RPMInstallCallback from mic.chroot import cleanup_mounts, setup_chrootenv, cleanup_chrootenv PATH_BOOTSTRAP = "/usr/sbin:/usr/bin:/sbin:/bin" RPMTRANS_FLAGS = [ rpm.RPMTRANS_FLAG_ALLFILES, rpm.RPMTRANS_FLAG_NOSCRIPTS, rpm.RPMTRANS_FLAG_NOTRIGGERS, ] RPMVSF_FLAGS = [ rpm._RPMVSF_NOSIGNATURES, rpm._RPMVSF_NODIGESTS ] RPMPROB_FLAGS = [ rpm.RPMPROB_FILTER_OLDPACKAGE, rpm.RPMPROB_FILTER_REPLACEPKG, rpm.RPMPROB_FILTER_IGNOREARCH ] class MiniBackend(object): def __init__(self, rootdir, arch=None, repomd=None): self._ts = None self.rootdir = os.path.abspath(rootdir) self.arch = arch self.repomd = repomd self.dlpkgs = [] self.localpkgs = {} self.optionals = [] self.preins = {} self.postins = {} self.scriptlets = False def __del__(self): try: del self.ts except: pass def get_ts(self): if not self._ts: self._ts = rpm.TransactionSet(self.rootdir) self._ts.setFlags(reduce(lambda x, y: x|y, RPMTRANS_FLAGS)) self._ts.setVSFlags(reduce(lambda x, y: x|y, RPMVSF_FLAGS)) self._ts.setProbFilter(reduce(lambda x, y: x|y, RPMPROB_FLAGS)) return self._ts def del_ts(self): if self._ts: self._ts.closeDB() self._ts = None ts = property(fget = lambda self: self.get_ts(), fdel = lambda self: self.del_ts(), doc="TransactionSet object") def selectPackage(self, pkg): if not pkg in self.dlpkgs: self.dlpkgs.append(pkg) def runInstall(self): # FIXME: check space self.downloadPkgs() self.installPkgs() if not self.scriptlets: return for pkg in self.preins.keys(): prog, script = self.preins[pkg] self.run_pkg_script(pkg, prog, script, '0') for pkg in self.postins.keys(): prog, script = self.postins[pkg] self.run_pkg_script(pkg, prog, script, '1') def downloadPkgs(self): nonexist = [] for pkg in self.dlpkgs: localpth = misc.get_package(pkg, self.repomd, self.arch) if localpth: self.localpkgs[pkg] = localpth elif pkg in self.optionals: # skip optional rpm continue else: # mark nonexist rpm nonexist.append(pkg) if nonexist: raise errors.BootstrapError("Can't get rpm binary: %s" % ','.join(nonexist)) def installPkgs(self): for pkg in self.localpkgs.keys(): rpmpath = self.localpkgs[pkg] hdr = readRpmHeader(self.ts, rpmpath) # save prein and postin scripts self.preins[pkg] = (hdr['PREINPROG'], hdr['PREIN']) self.postins[pkg] = (hdr['POSTINPROG'], hdr['POSTIN']) # mark pkg as install self.ts.addInstall(hdr, rpmpath, 'u') # run transaction self.ts.order() cb = RPMInstallCallback(self.ts) self.ts.run(cb.callback, '') def run_pkg_script(self, pkg, prog, script, arg): mychroot = lambda: os.chroot(self.rootdir) if not script: return if prog == "<lua>": prog = "/usr/bin/lua" tmpdir = os.path.join(self.rootdir, "tmp") if not os.path.exists(tmpdir): os.makedirs(tmpdir) tmpfd, tmpfp = tempfile.mkstemp(dir=tmpdir, prefix="%s.pre-" % pkg) script = script.replace('\r', '') os.write(tmpfd, script) os.close(tmpfd) os.chmod(tmpfp, 0700) try: script_fp = os.path.join('/tmp', os.path.basename(tmpfp)) subprocess.call([prog, script_fp, arg], preexec_fn=mychroot) except (OSError, IOError), err: msger.warning(str(err)) finally: os.unlink(tmpfp) class Bootstrap(object): def __init__(self, rootdir, distro, arch=None): self.rootdir = misc.mkdtemp(dir=rootdir, prefix=distro) self.distro = distro self.arch = arch self.logfile = None self.pkgslist = [] self.repomd = None def __del__(self): self.cleanup() def get_rootdir(self): if os.path.exists(self.rootdir): shutil.rmtree(self.rootdir, ignore_errors=True) os.makedirs(self.rootdir) return self.rootdir def dirsetup(self, rootdir=None): _path = lambda pth: os.path.join(rootdir, pth.lstrip('/')) if not rootdir: rootdir = self.rootdir try: # make /tmp and /etc path tmpdir = _path('/tmp') if not os.path.exists(tmpdir): os.makedirs(tmpdir) etcdir = _path('/etc') if not os.path.exists(etcdir): os.makedirs(etcdir) # touch distro file tzdist = _path('/etc/%s-release' % self.distro) if not os.path.exists(tzdist): with open(tzdist, 'w') as wf: wf.write("bootstrap") except: pass def create(self, repomd, pkglist, optlist=()): try: pkgmgr = MiniBackend(self.get_rootdir()) pkgmgr.arch = self.arch pkgmgr.repomd = repomd pkgmgr.optionals = list(optlist) map(pkgmgr.selectPackage, pkglist + list(optlist)) pkgmgr.runInstall() except (OSError, IOError, errors.CreatorError), err: raise errors.BootstrapError("%s" % err) def run(self, cmd, chdir, rootdir=None, bindmounts=None): def mychroot(): os.chroot(rootdir) os.chdir(chdir) def sync_timesetting(rootdir): try: # sync time and zone info to bootstrap if os.path.exists(rootdir + "/etc/localtime"): os.unlink(rootdir + "/etc/localtime") shutil.copyfile("/etc/localtime", rootdir + "/etc/localtime") except: pass def sync_passwdfile(rootdir): try: # sync passwd file to bootstrap, saving the user info if os.path.exists(rootdir + "/etc/passwd"): os.unlink(rootdir + "/etc/passwd") shutil.copyfile("/etc/passwd", rootdir + "/etc/passwd") except: pass if not rootdir: rootdir = self.rootdir if isinstance(cmd, list): shell = False else: shell = True env = os.environ env['PATH'] = "%s:%s" % (PATH_BOOTSTRAP, env['PATH']) retcode = 0 gloablmounts = None try: proxy.set_proxy_environ() gloablmounts = setup_chrootenv(rootdir, bindmounts, False) sync_timesetting(rootdir) sync_passwdfile(rootdir) retcode = subprocess.call(cmd, preexec_fn=mychroot, env=env, shell=shell) except (OSError, IOError): # add additional information to original exception value, tb = sys.exc_info()[1:] value = '%s: %s' % (value, ' '.join(cmd)) raise RuntimeError, value, tb finally: if self.logfile and os.path.isfile(self.logfile): msger.log(file(self.logfile).read()) cleanup_chrootenv(rootdir, bindmounts, gloablmounts) proxy.unset_proxy_environ() return retcode def cleanup(self): try: # clean mounts cleanup_mounts(self.rootdir) # remove rootdir shutil.rmtree(self.rootdir, ignore_errors=True) except: pass ########NEW FILE######## __FILENAME__ = chroot #!/usr/bin/python -tt # # Copyright (c) 2009, 2010, 2011 Intel, Inc. # # This program is free software; you can redistribute it and/or modify it # under the terms of the GNU General Public License as published by the Free # Software Foundation; version 2 of the License # # This program is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY # or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License # for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., 59 # Temple Place - Suite 330, Boston, MA 02111-1307, USA. from __future__ import with_statement import os import shutil import subprocess from mic import msger from mic.conf import configmgr from mic.utils import misc, errors, runner, fs_related chroot_lockfd = -1 chroot_lock = "" BIND_MOUNTS = ( "/proc", "/proc/sys/fs/binfmt_misc", "/sys", "/dev", "/dev/pts", "/dev/shm", "/var/lib/dbus", "/var/run/dbus", "/var/lock", ) def cleanup_after_chroot(targettype,imgmount,tmpdir,tmpmnt): if imgmount and targettype == "img": imgmount.cleanup() if tmpdir: shutil.rmtree(tmpdir, ignore_errors = True) if tmpmnt: shutil.rmtree(tmpmnt, ignore_errors = True) def check_bind_mounts(chrootdir, bindmounts): chrootmounts = [] for mount in bindmounts.split(";"): if not mount: continue srcdst = mount.split(":") if len(srcdst) == 1: srcdst.append("none") if not os.path.isdir(srcdst[0]): return False if srcdst[1] == "" or srcdst[1] == "none": srcdst[1] = None if srcdst[0] in BIND_MOUNTS or srcdst[0] == '/': continue if chrootdir: if not srcdst[1]: srcdst[1] = os.path.abspath(os.path.expanduser(srcdst[0])) else: srcdst[1] = os.path.abspath(os.path.expanduser(srcdst[1])) tmpdir = chrootdir + "/" + srcdst[1] if os.path.isdir(tmpdir): msger.warning("Warning: dir %s has existed." % tmpdir) return True def cleanup_mounts(chrootdir): umountcmd = misc.find_binary_path("umount") abs_chrootdir = os.path.abspath(chrootdir) mounts = open('/proc/mounts').readlines() for line in reversed(mounts): if abs_chrootdir not in line: continue point = line.split()[1] # '/' to avoid common name prefix if abs_chrootdir == point or point.startswith(abs_chrootdir + '/'): args = [ umountcmd, "-l", point ] ret = runner.quiet(args) if ret != 0: msger.warning("failed to unmount %s" % point) return 0 def setup_chrootenv(chrootdir, bindmounts = None, mountparent = True): global chroot_lockfd, chroot_lock def get_bind_mounts(chrootdir, bindmounts, mountparent = True): chrootmounts = [] if bindmounts in ("", None): bindmounts = "" for mount in bindmounts.split(";"): if not mount: continue srcdst = mount.split(":") srcdst[0] = os.path.abspath(os.path.expanduser(srcdst[0])) if len(srcdst) == 1: srcdst.append("none") # if some bindmount is not existed, but it's created inside # chroot, this is not expected if not os.path.exists(srcdst[0]): os.makedirs(srcdst[0]) if not os.path.isdir(srcdst[0]): continue if srcdst[0] in BIND_MOUNTS or srcdst[0] == '/': msger.verbose("%s will be mounted by default." % srcdst[0]) continue if srcdst[1] == "" or srcdst[1] == "none": srcdst[1] = None else: srcdst[1] = os.path.abspath(os.path.expanduser(srcdst[1])) if os.path.isdir(chrootdir + "/" + srcdst[1]): msger.warning("%s has existed in %s , skip it."\ % (srcdst[1], chrootdir)) continue chrootmounts.append(fs_related.BindChrootMount(srcdst[0], chrootdir, srcdst[1])) """Default bind mounts""" for pt in BIND_MOUNTS: if not os.path.exists(pt): continue chrootmounts.append(fs_related.BindChrootMount(pt, chrootdir, None)) if mountparent: chrootmounts.append(fs_related.BindChrootMount("/", chrootdir, "/parentroot", "ro")) for kernel in os.listdir("/lib/modules"): chrootmounts.append(fs_related.BindChrootMount( "/lib/modules/"+kernel, chrootdir, None, "ro")) return chrootmounts def bind_mount(chrootmounts): for b in chrootmounts: msger.verbose("bind_mount: %s -> %s" % (b.src, b.dest)) b.mount() def setup_resolv(chrootdir): try: shutil.copyfile("/etc/resolv.conf", chrootdir + "/etc/resolv.conf") except: pass globalmounts = get_bind_mounts(chrootdir, bindmounts, mountparent) bind_mount(globalmounts) setup_resolv(chrootdir) mtab = "/etc/mtab" dstmtab = chrootdir + mtab if not os.path.islink(dstmtab): shutil.copyfile(mtab, dstmtab) chroot_lock = os.path.join(chrootdir, ".chroot.lock") chroot_lockfd = open(chroot_lock, "w") return globalmounts def cleanup_chrootenv(chrootdir, bindmounts=None, globalmounts=()): global chroot_lockfd, chroot_lock def bind_unmount(chrootmounts): for b in reversed(chrootmounts): msger.verbose("bind_unmount: %s -> %s" % (b.src, b.dest)) b.unmount() def cleanup_resolv(chrootdir): try: fd = open(chrootdir + "/etc/resolv.conf", "w") fd.truncate(0) fd.close() except: pass def kill_processes(chrootdir): import glob for fp in glob.glob("/proc/*/root"): try: if os.readlink(fp) == chrootdir: pid = int(fp.split("/")[2]) os.kill(pid, 9) except: pass def cleanup_mountdir(chrootdir, bindmounts): if bindmounts == "" or bindmounts == None: return chrootmounts = [] for mount in bindmounts.split(";"): if not mount: continue srcdst = mount.split(":") if len(srcdst) == 1: srcdst.append("none") if srcdst[0] == "/": continue if srcdst[1] == "" or srcdst[1] == "none": srcdst[1] = srcdst[0] srcdst[1] = os.path.abspath(os.path.expanduser(srcdst[1])) tmpdir = chrootdir + "/" + srcdst[1] if os.path.isdir(tmpdir): if len(os.listdir(tmpdir)) == 0: shutil.rmtree(tmpdir, ignore_errors = True) else: msger.warning("Warning: dir %s isn't empty." % tmpdir) chroot_lockfd.close() bind_unmount(globalmounts) if not fs_related.my_fuser(chroot_lock): tmpdir = chrootdir + "/parentroot" if os.path.exists(tmpdir) and len(os.listdir(tmpdir)) == 0: shutil.rmtree(tmpdir, ignore_errors = True) cleanup_resolv(chrootdir) if os.path.exists(chrootdir + "/etc/mtab"): os.unlink(chrootdir + "/etc/mtab") kill_processes(chrootdir) cleanup_mountdir(chrootdir, bindmounts) def chroot(chrootdir, bindmounts = None, execute = "/bin/bash"): def mychroot(): os.chroot(chrootdir) os.chdir("/") if configmgr.chroot['saveto']: savefs = True saveto = configmgr.chroot['saveto'] wrnmsg = "Can't save chroot fs for dir %s exists" % saveto if saveto == chrootdir: savefs = False wrnmsg = "Dir %s is being used to chroot" % saveto elif os.path.exists(saveto): if msger.ask("Dir %s already exists, cleanup and continue?" % saveto): shutil.rmtree(saveto, ignore_errors = True) savefs = True else: savefs = False if savefs: msger.info("Saving image to directory %s" % saveto) fs_related.makedirs(os.path.dirname(os.path.abspath(saveto))) runner.quiet("cp -af %s %s" % (chrootdir, saveto)) devs = ['dev/fd', 'dev/stdin', 'dev/stdout', 'dev/stderr', 'etc/mtab'] ignlst = [os.path.join(saveto, x) for x in devs] map(os.unlink, filter(os.path.exists, ignlst)) else: msger.warning(wrnmsg) dev_null = os.open("/dev/null", os.O_WRONLY) files_to_check = ["/bin/bash", "/sbin/init"] architecture_found = False """ Register statically-linked qemu-arm if it is an ARM fs """ qemu_emulator = None for ftc in files_to_check: ftc = "%s/%s" % (chrootdir,ftc) # Return code of 'file' is "almost always" 0 based on some man pages # so we need to check the file existance first. if not os.path.exists(ftc): continue for line in runner.outs(['file', ftc]).splitlines(): if 'ARM' in line: qemu_emulator = misc.setup_qemu_emulator(chrootdir, "arm") architecture_found = True break if 'Intel' in line: architecture_found = True break if architecture_found: break os.close(dev_null) if not architecture_found: raise errors.CreatorError("Failed to get architecture from any of the " "following files %s from chroot." \ % files_to_check) try: msger.info("Launching shell. Exit to continue.\n" "----------------------------------") globalmounts = setup_chrootenv(chrootdir, bindmounts) subprocess.call(execute, preexec_fn = mychroot, shell=True) except OSError, err: raise errors.CreatorError("chroot err: %s" % str(err)) finally: cleanup_chrootenv(chrootdir, bindmounts, globalmounts) if qemu_emulator: os.unlink(chrootdir + qemu_emulator) ########NEW FILE######## __FILENAME__ = conf #!/usr/bin/python -tt # # Copyright (c) 2011 Intel, Inc. # # This program is free software; you can redistribute it and/or modify it # under the terms of the GNU General Public License as published by the Free # Software Foundation; version 2 of the License # # This program is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY # or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License # for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., 59 # Temple Place - Suite 330, Boston, MA 02111-1307, USA. import os, sys, re import ConfigParser from mic import msger from mic import kickstart from mic.utils import misc, runner, proxy, errors def get_siteconf(): mic_path = os.path.dirname(__file__) eos = mic_path.find('scripts') + len('scripts') scripts_path = mic_path[:eos] return scripts_path + "/lib/image/config/wic.conf" class ConfigMgr(object): prefer_backends = ["zypp", "yum"] DEFAULTS = {'common': { "distro_name": "Default Distribution", "plugin_dir": "/usr/lib/wic/plugins", # TODO use prefix also? }, 'create': { "tmpdir": '/var/tmp/wic', "cachedir": '/var/tmp/wic/cache', "outdir": './wic-output', "arch": None, # None means auto-detect "pkgmgr": "auto", "name": "output", "ksfile": None, "ks": None, "repomd": None, "local_pkgs_path": None, "release": None, "logfile": None, "record_pkgs": [], "pack_to": None, "name_prefix": None, "name_suffix": None, "proxy": None, "no_proxy": None, "copy_kernel": False, "install_pkgs": None, "repourl": {}, "localrepos": [], # save localrepos "runtime": "bootstrap", }, 'chroot': { "saveto": None, }, 'convert': { "shell": False, }, 'bootstrap': { "rootdir": '/var/tmp/wic-bootstrap', "packages": [], }, } # make the manager class as singleton _instance = None def __new__(cls, *args, **kwargs): if not cls._instance: cls._instance = super(ConfigMgr, cls).__new__(cls, *args, **kwargs) return cls._instance def __init__(self, ksconf=None, siteconf=None): # reset config options self.reset() if not siteconf: siteconf = get_siteconf() # initial options from siteconf self._siteconf = siteconf if ksconf: self._ksconf = ksconf def reset(self): self.__ksconf = None self.__siteconf = None # initialize the values with defaults for sec, vals in self.DEFAULTS.iteritems(): setattr(self, sec, vals) def __set_siteconf(self, siteconf): try: self.__siteconf = siteconf self._parse_siteconf(siteconf) except ConfigParser.Error, error: raise errors.ConfigError("%s" % error) def __get_siteconf(self): return self.__siteconf _siteconf = property(__get_siteconf, __set_siteconf) def __set_ksconf(self, ksconf): if not os.path.isfile(ksconf): msger.error('Cannot find ks file: %s' % ksconf) self.__ksconf = ksconf self._parse_kickstart(ksconf) def __get_ksconf(self): return self.__ksconf _ksconf = property(__get_ksconf, __set_ksconf) def _parse_siteconf(self, siteconf): if not siteconf: return if not os.path.exists(siteconf): msger.warning("cannot read config file: %s" % siteconf) return parser = ConfigParser.SafeConfigParser() parser.read(siteconf) for section in parser.sections(): if section in self.DEFAULTS: getattr(self, section).update(dict(parser.items(section))) # append common section items to other sections for section in self.DEFAULTS.keys(): if section != "common": getattr(self, section).update(self.common) # check and normalize the scheme of proxy url if self.create['proxy']: m = re.match('^(\w+)://.*', self.create['proxy']) if m: scheme = m.group(1) if scheme not in ('http', 'https', 'ftp', 'socks'): msger.error("%s: proxy scheme is incorrect" % siteconf) else: msger.warning("%s: proxy url w/o scheme, use http as default" % siteconf) self.create['proxy'] = "http://" + self.create['proxy'] proxy.set_proxies(self.create['proxy'], self.create['no_proxy']) # bootstrap option handling self.set_runtime(self.create['runtime']) if isinstance(self.bootstrap['packages'], basestring): packages = self.bootstrap['packages'].replace('\n', ' ') if packages.find(',') != -1: packages = packages.split(',') else: packages = packages.split() self.bootstrap['packages'] = packages def _parse_kickstart(self, ksconf=None): if not ksconf: return ksconf = misc.normalize_ksfile(ksconf, self.create['release'], self.create['arch']) ks = kickstart.read_kickstart(ksconf) self.create['ks'] = ks self.create['name'] = os.path.splitext(os.path.basename(ksconf))[0] self.create['name'] = misc.build_name(ksconf, self.create['release'], self.create['name_prefix'], self.create['name_suffix']) def set_runtime(self, runtime): if runtime not in ("bootstrap", "native"): msger.error("Invalid runtime mode: %s" % runtime) if misc.get_distro()[0] in ("tizen", "Tizen"): runtime = "native" self.create['runtime'] = runtime configmgr = ConfigMgr() ########NEW FILE######## __FILENAME__ = creator #!/usr/bin/python -tt # # Copyright (c) 2011 Intel, Inc. # # This program is free software; you can redistribute it and/or modify it # under the terms of the GNU General Public License as published by the Free # Software Foundation; version 2 of the License # # This program is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY # or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License # for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., 59 # Temple Place - Suite 330, Boston, MA 02111-1307, USA. import os, sys, re from optparse import SUPPRESS_HELP from mic import msger from mic.utils import cmdln, errors from mic.conf import configmgr from mic.plugin import pluginmgr class Creator(cmdln.Cmdln): """${name}: create an image Usage: ${name} SUBCOMMAND <ksfile> [OPTS] ${command_list} ${option_list} """ name = 'mic create(cr)' def __init__(self, *args, **kwargs): cmdln.Cmdln.__init__(self, *args, **kwargs) self._subcmds = [] # get cmds from pluginmgr # mix-in do_subcmd interface for subcmd, klass in pluginmgr.get_plugins('imager').iteritems(): if not hasattr(klass, 'do_create'): msger.warning("Unsurpport subcmd: %s" % subcmd) continue func = getattr(klass, 'do_create') setattr(self.__class__, "do_"+subcmd, func) self._subcmds.append(subcmd) def get_optparser(self): optparser = cmdln.CmdlnOptionParser(self) optparser.add_option('-d', '--debug', action='store_true', dest='debug', help=SUPPRESS_HELP) optparser.add_option('-v', '--verbose', action='store_true', dest='verbose', help=SUPPRESS_HELP) optparser.add_option('', '--logfile', type='string', dest='logfile', default=None, help='Path of logfile') optparser.add_option('-c', '--config', type='string', dest='config', default=None, help='Specify config file for mic') optparser.add_option('-k', '--cachedir', type='string', action='store', dest='cachedir', default=None, help='Cache directory to store the downloaded') optparser.add_option('-o', '--outdir', type='string', action='store', dest='outdir', default=None, help='Output directory') optparser.add_option('-A', '--arch', type='string', dest='arch', default=None, help='Specify repo architecture') optparser.add_option('', '--release', type='string', dest='release', default=None, metavar='RID', help='Generate a release of RID with all necessary' ' files, when @BUILD_ID@ is contained in ' 'kickstart file, it will be replaced by RID') optparser.add_option("", "--record-pkgs", type="string", dest="record_pkgs", default=None, help='Record the info of installed packages, ' 'multiple values can be specified which ' 'joined by ",", valid values: "name", ' '"content", "license", "vcs"') optparser.add_option('', '--pkgmgr', type='string', dest='pkgmgr', default=None, help='Specify backend package manager') optparser.add_option('', '--local-pkgs-path', type='string', dest='local_pkgs_path', default=None, help='Path for local pkgs(rpms) to be installed') optparser.add_option('', '--runtime', type='string', dest='runtime', default=None, help='Specify runtime mode, avaiable: bootstrap, native') # --taring-to is alias to --pack-to optparser.add_option('', '--taring-to', type='string', dest='pack_to', default=None, help=SUPPRESS_HELP) optparser.add_option('', '--pack-to', type='string', dest='pack_to', default=None, help='Pack the images together into the specified' ' achive, extension supported: .zip, .tar, ' '.tar.gz, .tar.bz2, etc. by default, .tar ' 'will be used') optparser.add_option('', '--copy-kernel', action='store_true', dest='copy_kernel', help='Copy kernel files from image /boot directory' ' to the image output directory.') optparser.add_option('', '--install-pkgs', type='string', action='store', dest='install_pkgs', default=None, help='Specify what type of packages to be installed,' ' valid: source, debuginfo, debugsource') optparser.add_option('', '--tmpfs', action='store_true', dest='enabletmpfs', help='Setup tmpdir as tmpfs to accelerate, experimental' ' feature, use it if you have more than 4G memory') optparser.add_option('', '--repourl', action='append', dest='repourl', default=[], help=SUPPRESS_HELP) return optparser def preoptparse(self, argv): optparser = self.get_optparser() largs = [] rargs = [] while argv: arg = argv.pop(0) if arg in ('-h', '--help'): rargs.append(arg) elif optparser.has_option(arg): largs.append(arg) if optparser.get_option(arg).takes_value(): try: largs.append(argv.pop(0)) except IndexError: raise errors.Usage("option %s requires arguments" % arg) else: if arg.startswith("--"): if "=" in arg: opt = arg.split("=")[0] else: opt = None elif arg.startswith("-") and len(arg) > 2: opt = arg[0:2] else: opt = None if opt and optparser.has_option(opt): largs.append(arg) else: rargs.append(arg) return largs + rargs def postoptparse(self): abspath = lambda pth: os.path.abspath(os.path.expanduser(pth)) if self.options.verbose: msger.set_loglevel('verbose') if self.options.debug: msger.set_loglevel('debug') if self.options.logfile: logfile_abs_path = abspath(self.options.logfile) if os.path.isdir(logfile_abs_path): raise errors.Usage("logfile's path %s should be file" % self.options.logfile) if not os.path.exists(os.path.dirname(logfile_abs_path)): os.makedirs(os.path.dirname(logfile_abs_path)) msger.set_interactive(False) msger.set_logfile(logfile_abs_path) configmgr.create['logfile'] = self.options.logfile if self.options.config: configmgr.reset() configmgr._siteconf = self.options.config if self.options.outdir is not None: configmgr.create['outdir'] = abspath(self.options.outdir) if self.options.cachedir is not None: configmgr.create['cachedir'] = abspath(self.options.cachedir) os.environ['ZYPP_LOCKFILE_ROOT'] = configmgr.create['cachedir'] for cdir in ('outdir', 'cachedir'): if os.path.exists(configmgr.create[cdir]) \ and not os.path.isdir(configmgr.create[cdir]): msger.error('Invalid directory specified: %s' \ % configmgr.create[cdir]) if self.options.local_pkgs_path is not None: if not os.path.exists(self.options.local_pkgs_path): msger.error('Local pkgs directory: \'%s\' not exist' \ % self.options.local_pkgs_path) configmgr.create['local_pkgs_path'] = self.options.local_pkgs_path if self.options.release: configmgr.create['release'] = self.options.release.rstrip('/') if self.options.record_pkgs: configmgr.create['record_pkgs'] = [] for infotype in self.options.record_pkgs.split(','): if infotype not in ('name', 'content', 'license', 'vcs'): raise errors.Usage('Invalid pkg recording: %s, valid ones:' ' "name", "content", "license", "vcs"' \ % infotype) configmgr.create['record_pkgs'].append(infotype) if self.options.arch is not None: supported_arch = sorted(rpmmisc.archPolicies.keys(), reverse=True) if self.options.arch in supported_arch: configmgr.create['arch'] = self.options.arch else: raise errors.Usage('Invalid architecture: "%s".\n' ' Supported architectures are: \n' ' %s' % (self.options.arch, ', '.join(supported_arch))) if self.options.pkgmgr is not None: configmgr.create['pkgmgr'] = self.options.pkgmgr if self.options.runtime: configmgr.set_runtime(self.options.runtime) if self.options.pack_to is not None: configmgr.create['pack_to'] = self.options.pack_to if self.options.copy_kernel: configmgr.create['copy_kernel'] = self.options.copy_kernel if self.options.install_pkgs: configmgr.create['install_pkgs'] = [] for pkgtype in self.options.install_pkgs.split(','): if pkgtype not in ('source', 'debuginfo', 'debugsource'): raise errors.Usage('Invalid parameter specified: "%s", ' 'valid values: source, debuginfo, ' 'debusource' % pkgtype) configmgr.create['install_pkgs'].append(pkgtype) if self.options.enabletmpfs: configmgr.create['enabletmpfs'] = self.options.enabletmpfs if self.options.repourl: for item in self.options.repourl: try: key, val = item.split('=') except: continue configmgr.create['repourl'][key] = val def main(self, argv=None): if argv is None: argv = sys.argv else: argv = argv[:] # don't modify caller's list self.optparser = self.get_optparser() if self.optparser: try: argv = self.preoptparse(argv) self.options, args = self.optparser.parse_args(argv) except cmdln.CmdlnUserError, ex: msg = "%s: %s\nTry '%s help' for info.\n"\ % (self.name, ex, self.name) msger.error(msg) except cmdln.StopOptionProcessing, ex: return 0 else: # optparser=None means no process for opts self.options, args = None, argv[1:] if not args: return self.emptyline() self.postoptparse() return self.cmd(args) def precmd(self, argv): # check help before cmd if '-h' in argv or '?' in argv or '--help' in argv or 'help' in argv: return argv if len(argv) == 1: return ['help', argv[0]] return argv def do_auto(self, subcmd, opts, *args): """${cmd_name}: auto detect image type from magic header Usage: ${name} ${cmd_name} <ksfile> ${cmd_option_list} """ def parse_magic_line(re_str, pstr, ptype='mic'): ptn = re.compile(re_str) m = ptn.match(pstr) if not m or not m.groups(): return None inline_argv = m.group(1).strip() if ptype == 'mic': m2 = re.search('(?P<format>\w+)', inline_argv) elif ptype == 'mic2': m2 = re.search('(-f|--format(=)?)\s*(?P<format>\w+)', inline_argv) else: return None if m2: cmdname = m2.group('format') inline_argv = inline_argv.replace(m2.group(0), '') return (cmdname, inline_argv) return None if len(args) != 1: raise errors.Usage("Extra arguments given") if not os.path.exists(args[0]): raise errors.CreatorError("Can't find the file: %s" % args[0]) with open(args[0], 'r') as rf: first_line = rf.readline() mic_re = '^#\s*-\*-mic-options-\*-\s+(.*)\s+-\*-mic-options-\*-' mic2_re = '^#\s*-\*-mic2-options-\*-\s+(.*)\s+-\*-mic2-options-\*-' result = parse_magic_line(mic_re, first_line, 'mic') \ or parse_magic_line(mic2_re, first_line, 'mic2') if not result: raise errors.KsError("Invalid magic line in file: %s" % args[0]) if result[0] not in self._subcmds: raise errors.KsError("Unsupport format '%s' in %s" % (result[0], args[0])) argv = ' '.join(result + args).split() self.main(argv) ########NEW FILE######## __FILENAME__ = baseimager #!/usr/bin/python -tt # # Copyright (c) 2007 Red Hat Inc. # Copyright (c) 2009, 2010, 2011 Intel, Inc. # # This program is free software; you can redistribute it and/or modify it # under the terms of the GNU General Public License as published by the Free # Software Foundation; version 2 of the License # # This program is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY # or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License # for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., 59 # Temple Place - Suite 330, Boston, MA 02111-1307, USA. from __future__ import with_statement import os, sys import stat import tempfile import shutil import subprocess import re import tarfile import glob from mic import kickstart from mic import msger from mic.utils.errors import CreatorError, Abort from mic.utils import misc, runner, fs_related as fs class BaseImageCreator(object): """Installs a system to a chroot directory. ImageCreator is the simplest creator class available; it will install and configure a system image according to the supplied kickstart file. e.g. import mic.imgcreate as imgcreate ks = imgcreate.read_kickstart("foo.ks") imgcreate.ImageCreator(ks, "foo").create() """ def __del__(self): self.cleanup() def __init__(self, createopts = None, pkgmgr = None): """Initialize an ImageCreator instance. ks -- a pykickstart.KickstartParser instance; this instance will be used to drive the install by e.g. providing the list of packages to be installed, the system configuration and %post scripts name -- a name for the image; used for e.g. image filenames or filesystem labels """ self.pkgmgr = pkgmgr self.__builddir = None self.__bindmounts = [] self.ks = None self.name = "target" self.tmpdir = "/var/tmp/wic" self.cachedir = "/var/tmp/wic/cache" self.workdir = "/var/tmp/wic/build" self.destdir = "." self.installerfw_prefix = "INSTALLERFW_" self.target_arch = "noarch" self._local_pkgs_path = None self.pack_to = None self.repourl = {} # If the kernel is save to the destdir when copy_kernel cmd is called. self._need_copy_kernel = False # setup tmpfs tmpdir when enabletmpfs is True self.enabletmpfs = False if createopts: # Mapping table for variables that have different names. optmap = {"pkgmgr" : "pkgmgr_name", "outdir" : "destdir", "arch" : "target_arch", "local_pkgs_path" : "_local_pkgs_path", "copy_kernel" : "_need_copy_kernel", } # update setting from createopts for key in createopts.keys(): if key in optmap: option = optmap[key] else: option = key setattr(self, option, createopts[key]) self.destdir = os.path.abspath(os.path.expanduser(self.destdir)) if 'release' in createopts and createopts['release']: self.name = createopts['release'] + '_' + self.name if self.pack_to: if '@NAME@' in self.pack_to: self.pack_to = self.pack_to.replace('@NAME@', self.name) (tar, ext) = os.path.splitext(self.pack_to) if ext in (".gz", ".bz2") and tar.endswith(".tar"): ext = ".tar" + ext if ext not in misc.pack_formats: self.pack_to += ".tar" self._dep_checks = ["ls", "bash", "cp", "echo", "modprobe"] # Output image file names self.outimage = [] # A flag to generate checksum self._genchecksum = False self._alt_initrd_name = None self._recording_pkgs = [] # available size in root fs, init to 0 self._root_fs_avail = 0 # Name of the disk image file that is created. self._img_name = None self.image_format = None # Save qemu emulator file name in order to clean up it finally self.qemu_emulator = None # No ks provided when called by convertor, so skip the dependency check if self.ks: # If we have btrfs partition we need to check necessary tools for part in self.ks.handler.partition.partitions: if part.fstype and part.fstype == "btrfs": self._dep_checks.append("mkfs.btrfs") break if self.target_arch and self.target_arch.startswith("arm"): for dep in self._dep_checks: if dep == "extlinux": self._dep_checks.remove(dep) if not os.path.exists("/usr/bin/qemu-arm") or \ not misc.is_statically_linked("/usr/bin/qemu-arm"): self._dep_checks.append("qemu-arm-static") if os.path.exists("/proc/sys/vm/vdso_enabled"): vdso_fh = open("/proc/sys/vm/vdso_enabled","r") vdso_value = vdso_fh.read().strip() vdso_fh.close() if (int)(vdso_value) == 1: msger.warning("vdso is enabled on your host, which might " "cause problems with arm emulations.\n" "\tYou can disable vdso with following command before " "starting image build:\n" "\techo 0 | sudo tee /proc/sys/vm/vdso_enabled") # make sure the specified tmpdir and cachedir exist if not os.path.exists(self.tmpdir): os.makedirs(self.tmpdir) if not os.path.exists(self.cachedir): os.makedirs(self.cachedir) # # Properties # def __get_instroot(self): if self.__builddir is None: raise CreatorError("_instroot is not valid before calling mount()") return self.__builddir + "/install_root" _instroot = property(__get_instroot) """The location of the install root directory. This is the directory into which the system is installed. Subclasses may mount a filesystem image here or copy files to/from here. Note, this directory does not exist before ImageCreator.mount() is called. Note also, this is a read-only attribute. """ def __get_outdir(self): if self.__builddir is None: raise CreatorError("_outdir is not valid before calling mount()") return self.__builddir + "/out" _outdir = property(__get_outdir) """The staging location for the final image. This is where subclasses should stage any files that are part of the final image. ImageCreator.package() will copy any files found here into the requested destination directory. Note, this directory does not exist before ImageCreator.mount() is called. Note also, this is a read-only attribute. """ # # Hooks for subclasses # def _mount_instroot(self, base_on = None): """Mount or prepare the install root directory. This is the hook where subclasses may prepare the install root by e.g. mounting creating and loopback mounting a filesystem image to _instroot. There is no default implementation. base_on -- this is the value passed to mount() and can be interpreted as the subclass wishes; it might e.g. be the location of a previously created ISO containing a system image. """ pass def _unmount_instroot(self): """Undo anything performed in _mount_instroot(). This is the hook where subclasses must undo anything which was done in _mount_instroot(). For example, if a filesystem image was mounted onto _instroot, it should be unmounted here. There is no default implementation. """ pass def _create_bootconfig(self): """Configure the image so that it's bootable. This is the hook where subclasses may prepare the image for booting by e.g. creating an initramfs and bootloader configuration. This hook is called while the install root is still mounted, after the packages have been installed and the kickstart configuration has been applied, but before the %post scripts have been executed. There is no default implementation. """ pass def _stage_final_image(self): """Stage the final system image in _outdir. This is the hook where subclasses should place the image in _outdir so that package() can copy it to the requested destination directory. By default, this moves the install root into _outdir. """ shutil.move(self._instroot, self._outdir + "/" + self.name) def get_installed_packages(self): return self._pkgs_content.keys() def _save_recording_pkgs(self, destdir): """Save the list or content of installed packages to file. """ pkgs = self._pkgs_content.keys() pkgs.sort() # inplace op if not os.path.exists(destdir): os.makedirs(destdir) content = None if 'vcs' in self._recording_pkgs: vcslst = ["%s %s" % (k, v) for (k, v) in self._pkgs_vcsinfo.items()] content = '\n'.join(sorted(vcslst)) elif 'name' in self._recording_pkgs: content = '\n'.join(pkgs) if content: namefile = os.path.join(destdir, self.name + '.packages') f = open(namefile, "w") f.write(content) f.close() self.outimage.append(namefile); # if 'content', save more details if 'content' in self._recording_pkgs: contfile = os.path.join(destdir, self.name + '.files') f = open(contfile, "w") for pkg in pkgs: content = pkg + '\n' pkgcont = self._pkgs_content[pkg] content += ' ' content += '\n '.join(pkgcont) content += '\n' content += '\n' f.write(content) f.close() self.outimage.append(contfile) if 'license' in self._recording_pkgs: licensefile = os.path.join(destdir, self.name + '.license') f = open(licensefile, "w") f.write('Summary:\n') for license in reversed(sorted(self._pkgs_license, key=\ lambda license: len(self._pkgs_license[license]))): f.write(" - %s: %s\n" \ % (license, len(self._pkgs_license[license]))) f.write('\nDetails:\n') for license in reversed(sorted(self._pkgs_license, key=\ lambda license: len(self._pkgs_license[license]))): f.write(" - %s:\n" % (license)) for pkg in sorted(self._pkgs_license[license]): f.write(" - %s\n" % (pkg)) f.write('\n') f.close() self.outimage.append(licensefile) def _get_required_packages(self): """Return a list of required packages. This is the hook where subclasses may specify a set of packages which it requires to be installed. This returns an empty list by default. Note, subclasses should usually chain up to the base class implementation of this hook. """ return [] def _get_excluded_packages(self): """Return a list of excluded packages. This is the hook where subclasses may specify a set of packages which it requires _not_ to be installed. This returns an empty list by default. Note, subclasses should usually chain up to the base class implementation of this hook. """ return [] def _get_local_packages(self): """Return a list of rpm path to be local installed. This is the hook where subclasses may specify a set of rpms which it requires to be installed locally. This returns an empty list by default. Note, subclasses should usually chain up to the base class implementation of this hook. """ if self._local_pkgs_path: if os.path.isdir(self._local_pkgs_path): return glob.glob( os.path.join(self._local_pkgs_path, '*.rpm')) elif os.path.splitext(self._local_pkgs_path)[-1] == '.rpm': return [self._local_pkgs_path] return [] def _get_fstab(self): """Return the desired contents of /etc/fstab. This is the hook where subclasses may specify the contents of /etc/fstab by returning a string containing the desired contents. A sensible default implementation is provided. """ s = "/dev/root / %s %s 0 0\n" \ % (self._fstype, "defaults,noatime" if not self._fsopts else self._fsopts) s += self._get_fstab_special() return s def _get_fstab_special(self): s = "devpts /dev/pts devpts gid=5,mode=620 0 0\n" s += "tmpfs /dev/shm tmpfs defaults 0 0\n" s += "proc /proc proc defaults 0 0\n" s += "sysfs /sys sysfs defaults 0 0\n" return s def _set_part_env(self, pnum, prop, value): """ This is a helper function which generates an environment variable for a property "prop" with value "value" of a partition number "pnum". The naming convention is: * Variables start with INSTALLERFW_PART * Then goes the partition number, the order is the same as specified in the KS file * Then goes the property name """ if value == None: value = "" else: value = str(value) name = self.installerfw_prefix + ("PART%d_" % pnum) + prop return { name : value } def _get_post_scripts_env(self, in_chroot): """Return an environment dict for %post scripts. This is the hook where subclasses may specify some environment variables for %post scripts by return a dict containing the desired environment. in_chroot -- whether this %post script is to be executed chroot()ed into _instroot. """ env = {} pnum = 0 for p in kickstart.get_partitions(self.ks): env.update(self._set_part_env(pnum, "SIZE", p.size)) env.update(self._set_part_env(pnum, "MOUNTPOINT", p.mountpoint)) env.update(self._set_part_env(pnum, "FSTYPE", p.fstype)) env.update(self._set_part_env(pnum, "LABEL", p.label)) env.update(self._set_part_env(pnum, "FSOPTS", p.fsopts)) env.update(self._set_part_env(pnum, "BOOTFLAG", p.active)) env.update(self._set_part_env(pnum, "ALIGN", p.align)) env.update(self._set_part_env(pnum, "TYPE_ID", p.part_type)) env.update(self._set_part_env(pnum, "DEVNODE", "/dev/%s%d" % (p.disk, pnum + 1))) pnum += 1 # Count of paritions env[self.installerfw_prefix + "PART_COUNT"] = str(pnum) # Partition table format ptable_format = self.ks.handler.bootloader.ptable env[self.installerfw_prefix + "PTABLE_FORMAT"] = ptable_format # The kerned boot parameters kernel_opts = self.ks.handler.bootloader.appendLine env[self.installerfw_prefix + "KERNEL_OPTS"] = kernel_opts # Name of the distribution env[self.installerfw_prefix + "DISTRO_NAME"] = self.distro_name # Name of the image creation tool env[self.installerfw_prefix + "INSTALLER_NAME"] = "wic" # The real current location of the mounted file-systems if in_chroot: mount_prefix = "/" else: mount_prefix = self._instroot env[self.installerfw_prefix + "MOUNT_PREFIX"] = mount_prefix # These are historical variables which lack the common name prefix if not in_chroot: env["INSTALL_ROOT"] = self._instroot env["IMG_NAME"] = self._name return env def __get_imgname(self): return self.name _name = property(__get_imgname) """The name of the image file. """ def _get_kernel_versions(self): """Return a dict detailing the available kernel types/versions. This is the hook where subclasses may override what kernel types and versions should be available for e.g. creating the booloader configuration. A dict should be returned mapping the available kernel types to a list of the available versions for those kernels. The default implementation uses rpm to iterate over everything providing 'kernel', finds /boot/vmlinuz-* and returns the version obtained from the vmlinuz filename. (This can differ from the kernel RPM's n-v-r in the case of e.g. xen) """ def get_kernel_versions(instroot): ret = {} versions = set() files = glob.glob(instroot + "/boot/vmlinuz-*") for file in files: version = os.path.basename(file)[8:] if version is None: continue versions.add(version) ret["kernel"] = list(versions) return ret def get_version(header): version = None for f in header['filenames']: if f.startswith('/boot/vmlinuz-'): version = f[14:] return version if self.ks is None: return get_kernel_versions(self._instroot) ts = rpm.TransactionSet(self._instroot) ret = {} for header in ts.dbMatch('provides', 'kernel'): version = get_version(header) if version is None: continue name = header['name'] if not name in ret: ret[name] = [version] elif not version in ret[name]: ret[name].append(version) return ret # # Helpers for subclasses # def _do_bindmounts(self): """Mount various system directories onto _instroot. This method is called by mount(), but may also be used by subclasses in order to re-mount the bindmounts after modifying the underlying filesystem. """ for b in self.__bindmounts: b.mount() def _undo_bindmounts(self): """Unmount the bind-mounted system directories from _instroot. This method is usually only called by unmount(), but may also be used by subclasses in order to gain access to the filesystem obscured by the bindmounts - e.g. in order to create device nodes on the image filesystem. """ self.__bindmounts.reverse() for b in self.__bindmounts: b.unmount() def _chroot(self): """Chroot into the install root. This method may be used by subclasses when executing programs inside the install root e.g. subprocess.call(["/bin/ls"], preexec_fn = self.chroot) """ os.chroot(self._instroot) os.chdir("/") def _mkdtemp(self, prefix = "tmp-"): """Create a temporary directory. This method may be used by subclasses to create a temporary directory for use in building the final image - e.g. a subclass might create a temporary directory in order to bundle a set of files into a package. The subclass may delete this directory if it wishes, but it will be automatically deleted by cleanup(). The absolute path to the temporary directory is returned. Note, this method should only be called after mount() has been called. prefix -- a prefix which should be used when creating the directory; defaults to "tmp-". """ self.__ensure_builddir() return tempfile.mkdtemp(dir = self.__builddir, prefix = prefix) def _mkstemp(self, prefix = "tmp-"): """Create a temporary file. This method may be used by subclasses to create a temporary file for use in building the final image - e.g. a subclass might need a temporary location to unpack a compressed file. The subclass may delete this file if it wishes, but it will be automatically deleted by cleanup(). A tuple containing a file descriptor (returned from os.open() and the absolute path to the temporary directory is returned. Note, this method should only be called after mount() has been called. prefix -- a prefix which should be used when creating the file; defaults to "tmp-". """ self.__ensure_builddir() return tempfile.mkstemp(dir = self.__builddir, prefix = prefix) def _mktemp(self, prefix = "tmp-"): """Create a temporary file. This method simply calls _mkstemp() and closes the returned file descriptor. The absolute path to the temporary file is returned. Note, this method should only be called after mount() has been called. prefix -- a prefix which should be used when creating the file; defaults to "tmp-". """ (f, path) = self._mkstemp(prefix) os.close(f) return path # # Actual implementation # def __ensure_builddir(self): if not self.__builddir is None: return try: self.workdir = os.path.join(self.tmpdir, "build") if not os.path.exists(self.workdir): os.makedirs(self.workdir) self.__builddir = tempfile.mkdtemp(dir = self.workdir, prefix = "imgcreate-") except OSError, (err, msg): raise CreatorError("Failed create build directory in %s: %s" % (self.tmpdir, msg)) def get_cachedir(self, cachedir = None): if self.cachedir: return self.cachedir self.__ensure_builddir() if cachedir: self.cachedir = cachedir else: self.cachedir = self.__builddir + "/wic-cache" fs.makedirs(self.cachedir) return self.cachedir def __sanity_check(self): """Ensure that the config we've been given is sane.""" if not (kickstart.get_packages(self.ks) or kickstart.get_groups(self.ks)): raise CreatorError("No packages or groups specified") kickstart.convert_method_to_repo(self.ks) if not kickstart.get_repos(self.ks): raise CreatorError("No repositories specified") def __write_fstab(self): fstab_contents = self._get_fstab() if fstab_contents: fstab = open(self._instroot + "/etc/fstab", "w") fstab.write(fstab_contents) fstab.close() def __create_minimal_dev(self): """Create a minimal /dev so that we don't corrupt the host /dev""" origumask = os.umask(0000) devices = (('null', 1, 3, 0666), ('urandom',1, 9, 0666), ('random', 1, 8, 0666), ('full', 1, 7, 0666), ('ptmx', 5, 2, 0666), ('tty', 5, 0, 0666), ('zero', 1, 5, 0666)) links = (("/proc/self/fd", "/dev/fd"), ("/proc/self/fd/0", "/dev/stdin"), ("/proc/self/fd/1", "/dev/stdout"), ("/proc/self/fd/2", "/dev/stderr")) for (node, major, minor, perm) in devices: if not os.path.exists(self._instroot + "/dev/" + node): os.mknod(self._instroot + "/dev/" + node, perm | stat.S_IFCHR, os.makedev(major,minor)) for (src, dest) in links: if not os.path.exists(self._instroot + dest): os.symlink(src, self._instroot + dest) os.umask(origumask) def __setup_tmpdir(self): if not self.enabletmpfs: return runner.show('mount -t tmpfs -o size=4G tmpfs %s' % self.workdir) def __clean_tmpdir(self): if not self.enabletmpfs: return runner.show('umount -l %s' % self.workdir) def mount(self, base_on = None, cachedir = None): """Setup the target filesystem in preparation for an install. This function sets up the filesystem which the ImageCreator will install into and configure. The ImageCreator class merely creates an install root directory, bind mounts some system directories (e.g. /dev) and writes out /etc/fstab. Other subclasses may also e.g. create a sparse file, format it and loopback mount it to the install root. base_on -- a previous install on which to base this install; defaults to None, causing a new image to be created cachedir -- a directory in which to store the Yum cache; defaults to None, causing a new cache to be created; by setting this to another directory, the same cache can be reused across multiple installs. """ self.__setup_tmpdir() self.__ensure_builddir() self._mount_instroot(base_on) def unmount(self): """Unmounts the target filesystem. The ImageCreator class detaches the system from the install root, but other subclasses may also detach the loopback mounted filesystem image from the install root. """ self._unmount_instroot() def cleanup(self): """Unmounts the target filesystem and deletes temporary files. This method calls unmount() and then deletes any temporary files and directories that were created on the host system while building the image. Note, make sure to call this method once finished with the creator instance in order to ensure no stale files are left on the host e.g.: creator = ImageCreator(ks, name) try: creator.create() finally: creator.cleanup() """ if not self.__builddir: return self.unmount() shutil.rmtree(self.__builddir, ignore_errors = True) self.__builddir = None self.__clean_tmpdir() def __is_excluded_pkg(self, pkg): if pkg in self._excluded_pkgs: self._excluded_pkgs.remove(pkg) return True for xpkg in self._excluded_pkgs: if xpkg.endswith('*'): if pkg.startswith(xpkg[:-1]): return True elif xpkg.startswith('*'): if pkg.endswith(xpkg[1:]): return True return None def __select_packages(self, pkg_manager): skipped_pkgs = [] for pkg in self._required_pkgs: e = pkg_manager.selectPackage(pkg) if e: if kickstart.ignore_missing(self.ks): skipped_pkgs.append(pkg) elif self.__is_excluded_pkg(pkg): skipped_pkgs.append(pkg) else: raise CreatorError("Failed to find package '%s' : %s" % (pkg, e)) for pkg in skipped_pkgs: msger.warning("Skipping missing package '%s'" % (pkg,)) def __select_groups(self, pkg_manager): skipped_groups = [] for group in self._required_groups: e = pkg_manager.selectGroup(group.name, group.include) if e: if kickstart.ignore_missing(self.ks): skipped_groups.append(group) else: raise CreatorError("Failed to find group '%s' : %s" % (group.name, e)) for group in skipped_groups: msger.warning("Skipping missing group '%s'" % (group.name,)) def __deselect_packages(self, pkg_manager): for pkg in self._excluded_pkgs: pkg_manager.deselectPackage(pkg) def __localinst_packages(self, pkg_manager): for rpm_path in self._get_local_packages(): pkg_manager.installLocal(rpm_path) def __preinstall_packages(self, pkg_manager): if not self.ks: return self._preinstall_pkgs = kickstart.get_pre_packages(self.ks) for pkg in self._preinstall_pkgs: pkg_manager.preInstall(pkg) def __attachment_packages(self, pkg_manager): if not self.ks: return self._attachment = [] for item in kickstart.get_attachment(self.ks): if item.startswith('/'): fpaths = os.path.join(self._instroot, item.lstrip('/')) for fpath in glob.glob(fpaths): self._attachment.append(fpath) continue filelist = pkg_manager.getFilelist(item) if filelist: # found rpm in rootfs for pfile in pkg_manager.getFilelist(item): fpath = os.path.join(self._instroot, pfile.lstrip('/')) self._attachment.append(fpath) continue # try to retrieve rpm file (url, proxies) = pkg_manager.package_url(item) if not url: msger.warning("Can't get url from repo for %s" % item) continue fpath = os.path.join(self.cachedir, os.path.basename(url)) if not os.path.exists(fpath): # download pkgs try: fpath = grabber.myurlgrab(url, fpath, proxies, None) except CreatorError: raise tmpdir = self._mkdtemp() misc.extract_rpm(fpath, tmpdir) for (root, dirs, files) in os.walk(tmpdir): for fname in files: fpath = os.path.join(root, fname) self._attachment.append(fpath) def install(self, repo_urls=None): """Install packages into the install root. This function installs the packages listed in the supplied kickstart into the install root. By default, the packages are installed from the repository URLs specified in the kickstart. repo_urls -- a dict which maps a repository name to a repository URL; if supplied, this causes any repository URLs specified in the kickstart to be overridden. """ # initialize pkg list to install if self.ks: self.__sanity_check() self._required_pkgs = \ kickstart.get_packages(self.ks, self._get_required_packages()) self._excluded_pkgs = \ kickstart.get_excluded(self.ks, self._get_excluded_packages()) self._required_groups = kickstart.get_groups(self.ks) else: self._required_pkgs = None self._excluded_pkgs = None self._required_groups = None pkg_manager = self.get_pkg_manager() pkg_manager.setup() if hasattr(self, 'install_pkgs') and self.install_pkgs: if 'debuginfo' in self.install_pkgs: pkg_manager.install_debuginfo = True for repo in kickstart.get_repos(self.ks, repo_urls): (name, baseurl, mirrorlist, inc, exc, proxy, proxy_username, proxy_password, debuginfo, source, gpgkey, disable, ssl_verify, nocache, cost, priority) = repo yr = pkg_manager.addRepository(name, baseurl, mirrorlist, proxy, proxy_username, proxy_password, inc, exc, ssl_verify, nocache, cost, priority) if kickstart.exclude_docs(self.ks): rpm.addMacro("_excludedocs", "1") rpm.addMacro("_dbpath", "/var/lib/rpm") rpm.addMacro("__file_context_path", "%{nil}") if kickstart.inst_langs(self.ks) != None: rpm.addMacro("_install_langs", kickstart.inst_langs(self.ks)) try: self.__preinstall_packages(pkg_manager) self.__select_packages(pkg_manager) self.__select_groups(pkg_manager) self.__deselect_packages(pkg_manager) self.__localinst_packages(pkg_manager) BOOT_SAFEGUARD = 256L * 1024 * 1024 # 256M checksize = self._root_fs_avail if checksize: checksize -= BOOT_SAFEGUARD if self.target_arch: pkg_manager._add_prob_flags(rpm.RPMPROB_FILTER_IGNOREARCH) pkg_manager.runInstall(checksize) except CreatorError, e: raise except KeyboardInterrupt: raise else: self._pkgs_content = pkg_manager.getAllContent() self._pkgs_license = pkg_manager.getPkgsLicense() self._pkgs_vcsinfo = pkg_manager.getVcsInfo() self.__attachment_packages(pkg_manager) finally: pkg_manager.close() # hook post install self.postinstall() # do some clean up to avoid lvm info leakage. this sucks. for subdir in ("cache", "backup", "archive"): lvmdir = self._instroot + "/etc/lvm/" + subdir try: for f in os.listdir(lvmdir): os.unlink(lvmdir + "/" + f) except: pass def postinstall(self): self.copy_attachment() def __run_post_scripts(self): msger.info("Running scripts ...") if os.path.exists(self._instroot + "/tmp"): shutil.rmtree(self._instroot + "/tmp") os.mkdir (self._instroot + "/tmp", 0755) for s in kickstart.get_post_scripts(self.ks): (fd, path) = tempfile.mkstemp(prefix = "ks-script-", dir = self._instroot + "/tmp") s.script = s.script.replace("\r", "") os.write(fd, s.script) os.close(fd) os.chmod(path, 0700) env = self._get_post_scripts_env(s.inChroot) if not s.inChroot: preexec = None script = path else: preexec = self._chroot script = "/tmp/" + os.path.basename(path) try: try: subprocess.call([s.interp, script], preexec_fn = preexec, env = env, stdout = sys.stdout, stderr = sys.stderr) except OSError, (err, msg): raise CreatorError("Failed to execute %%post script " "with '%s' : %s" % (s.interp, msg)) finally: os.unlink(path) def __save_repo_keys(self, repodata): if not repodata: return None gpgkeydir = "/etc/pki/rpm-gpg" fs.makedirs(self._instroot + gpgkeydir) for repo in repodata: if repo["repokey"]: repokey = gpgkeydir + "/RPM-GPG-KEY-%s" % repo["name"] shutil.copy(repo["repokey"], self._instroot + repokey) def configure(self, repodata = None): """Configure the system image according to the kickstart. This method applies the (e.g. keyboard or network) configuration specified in the kickstart and executes the kickstart %post scripts. If necessary, it also prepares the image to be bootable by e.g. creating an initrd and bootloader configuration. """ ksh = self.ks.handler msger.info('Applying configurations ...') try: kickstart.LanguageConfig(self._instroot).apply(ksh.lang) kickstart.KeyboardConfig(self._instroot).apply(ksh.keyboard) kickstart.TimezoneConfig(self._instroot).apply(ksh.timezone) #kickstart.AuthConfig(self._instroot).apply(ksh.authconfig) kickstart.FirewallConfig(self._instroot).apply(ksh.firewall) kickstart.RootPasswordConfig(self._instroot).apply(ksh.rootpw) kickstart.UserConfig(self._instroot).apply(ksh.user) kickstart.ServicesConfig(self._instroot).apply(ksh.services) kickstart.XConfig(self._instroot).apply(ksh.xconfig) kickstart.NetworkConfig(self._instroot).apply(ksh.network) kickstart.RPMMacroConfig(self._instroot).apply(self.ks) kickstart.DesktopConfig(self._instroot).apply(ksh.desktop) self.__save_repo_keys(repodata) kickstart.MoblinRepoConfig(self._instroot).apply(ksh.repo, repodata, self.repourl) except: msger.warning("Failed to apply configuration to image") raise self._create_bootconfig() self.__run_post_scripts() def launch_shell(self, launch): """Launch a shell in the install root. This method is launches a bash shell chroot()ed in the install root; this can be useful for debugging. """ if launch: msger.info("Launching shell. Exit to continue.") subprocess.call(["/bin/bash"], preexec_fn = self._chroot) def do_genchecksum(self, image_name): if not self._genchecksum: return md5sum = misc.get_md5sum(image_name) with open(image_name + ".md5sum", "w") as f: f.write("%s %s" % (md5sum, os.path.basename(image_name))) self.outimage.append(image_name+".md5sum") def package(self, destdir = "."): """Prepares the created image for final delivery. In its simplest form, this method merely copies the install root to the supplied destination directory; other subclasses may choose to package the image by e.g. creating a bootable ISO containing the image and bootloader configuration. destdir -- the directory into which the final image should be moved; this defaults to the current directory. """ self._stage_final_image() if not os.path.exists(destdir): fs.makedirs(destdir) if self._recording_pkgs: self._save_recording_pkgs(destdir) # For image formats with two or multiple image files, it will be # better to put them under a directory if self.image_format in ("raw", "vmdk", "vdi", "nand", "mrstnand"): destdir = os.path.join(destdir, "%s-%s" \ % (self.name, self.image_format)) msger.debug("creating destination dir: %s" % destdir) fs.makedirs(destdir) # Ensure all data is flushed to _outdir runner.quiet('sync') misc.check_space_pre_cp(self._outdir, destdir) for f in os.listdir(self._outdir): shutil.move(os.path.join(self._outdir, f), os.path.join(destdir, f)) self.outimage.append(os.path.join(destdir, f)) self.do_genchecksum(os.path.join(destdir, f)) def print_outimage_info(self): msg = "The new image can be found here:\n" self.outimage.sort() for file in self.outimage: msg += ' %s\n' % os.path.abspath(file) msger.info(msg) def check_depend_tools(self): for tool in self._dep_checks: fs.find_binary_path(tool) def package_output(self, image_format, destdir = ".", package="none"): if not package or package == "none": return destdir = os.path.abspath(os.path.expanduser(destdir)) (pkg, comp) = os.path.splitext(package) if comp: comp=comp.lstrip(".") if pkg == "tar": if comp: dst = "%s/%s-%s.tar.%s" %\ (destdir, self.name, image_format, comp) else: dst = "%s/%s-%s.tar" %\ (destdir, self.name, image_format) msger.info("creating %s" % dst) tar = tarfile.open(dst, "w:" + comp) for file in self.outimage: msger.info("adding %s to %s" % (file, dst)) tar.add(file, arcname=os.path.join("%s-%s" \ % (self.name, image_format), os.path.basename(file))) if os.path.isdir(file): shutil.rmtree(file, ignore_errors = True) else: os.remove(file) tar.close() '''All the file in outimage has been packaged into tar.* file''' self.outimage = [dst] def release_output(self, config, destdir, release): """ Create release directory and files """ def _rpath(fn): """ release path """ return os.path.join(destdir, fn) outimages = self.outimage # new ks new_kspath = _rpath(self.name+'.ks') with open(config) as fr: with open(new_kspath, "w") as wf: # When building a release we want to make sure the .ks # file generates the same build even when --release not used. wf.write(fr.read().replace("@BUILD_ID@", release)) outimages.append(new_kspath) # save log file, logfile is only available in creator attrs if hasattr(self, 'logfile') and not self.logfile: log_path = _rpath(self.name + ".log") # touch the log file, else outimages will filter it out with open(log_path, 'w') as wf: wf.write('') msger.set_logfile(log_path) outimages.append(_rpath(self.name + ".log")) # rename iso and usbimg for f in os.listdir(destdir): if f.endswith(".iso"): newf = f[:-4] + '.img' elif f.endswith(".usbimg"): newf = f[:-7] + '.img' else: continue os.rename(_rpath(f), _rpath(newf)) outimages.append(_rpath(newf)) # generate MD5SUMS with open(_rpath("MD5SUMS"), "w") as wf: for f in os.listdir(destdir): if f == "MD5SUMS": continue if os.path.isdir(os.path.join(destdir, f)): continue md5sum = misc.get_md5sum(_rpath(f)) # There needs to be two spaces between the sum and # filepath to match the syntax with md5sum. # This way also md5sum -c MD5SUMS can be used by users wf.write("%s *%s\n" % (md5sum, f)) outimages.append("%s/MD5SUMS" % destdir) # Filter out the nonexist file for fp in outimages[:]: if not os.path.exists("%s" % fp): outimages.remove(fp) def copy_kernel(self): """ Copy kernel files to the outimage directory. NOTE: This needs to be called before unmounting the instroot. """ if not self._need_copy_kernel: return if not os.path.exists(self.destdir): os.makedirs(self.destdir) for kernel in glob.glob("%s/boot/vmlinuz-*" % self._instroot): kernelfilename = "%s/%s-%s" % (self.destdir, self.name, os.path.basename(kernel)) msger.info('copy kernel file %s as %s' % (os.path.basename(kernel), kernelfilename)) shutil.copy(kernel, kernelfilename) self.outimage.append(kernelfilename) def copy_attachment(self): """ Subclass implement it to handle attachment files NOTE: This needs to be called before unmounting the instroot. """ pass def get_pkg_manager(self): return self.pkgmgr(target_arch = self.target_arch, instroot = self._instroot, cachedir = self.cachedir) ########NEW FILE######## __FILENAME__ = direct # ex:ts=4:sw=4:sts=4:et # -*- tab-width: 4; c-basic-offset: 4; indent-tabs-mode: nil -*- # # Copyright (c) 2013, Intel Corporation. # All rights reserved. # # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License version 2 as # published by the Free Software Foundation. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., # 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA. # # DESCRIPTION # This implements the 'direct' image creator class for 'wic', based # loosely on the raw image creator from 'mic' # # AUTHORS # Tom Zanussi <tom.zanussi (at] linux.intel.com> # import os import stat import shutil from mic import kickstart, msger from mic.utils import fs_related, runner, misc from mic.utils.partitionedfs import PartitionedMount from mic.utils.errors import CreatorError, MountError from mic.imager.baseimager import BaseImageCreator from mic.utils.oe.misc import * from mic.plugin import pluginmgr disk_methods = { "do_install_disk":None, } class DirectImageCreator(BaseImageCreator): """ Installs a system into a file containing a partitioned disk image. DirectImageCreator is an advanced ImageCreator subclass; an image file is formatted with a partition table, each partition created from a rootfs or other OpenEmbedded build artifact and dd'ed into the virtual disk. The disk image can subsequently be dd'ed onto media and used on actual hardware. """ def __init__(self, oe_builddir, image_output_dir, rootfs_dir, bootimg_dir, kernel_dir, native_sysroot, hdddir, staging_data_dir, creatoropts=None, pkgmgr=None, compress_image=None, generate_bmap=None, fstab_entry="uuid"): """ Initialize a DirectImageCreator instance. This method takes the same arguments as ImageCreator.__init__() """ BaseImageCreator.__init__(self, creatoropts, pkgmgr) self.__instimage = None self.__imgdir = None self.__disks = {} self.__disk_format = "direct" self._disk_names = [] self._ptable_format = self.ks.handler.bootloader.ptable self.use_uuid = fstab_entry == "uuid" self.compress_image = compress_image self.bmap_needed = generate_bmap self.oe_builddir = oe_builddir if image_output_dir: self.tmpdir = image_output_dir self.cachedir = "%s/cache" % image_output_dir self.rootfs_dir = rootfs_dir self.bootimg_dir = bootimg_dir self.kernel_dir = kernel_dir self.native_sysroot = native_sysroot self.hdddir = hdddir self.staging_data_dir = staging_data_dir def __write_fstab(self, image_rootfs): """overriden to generate fstab (temporarily) in rootfs. This is called from mount_instroot, make sure it doesn't get called from BaseImage.mount()""" if image_rootfs is None: return None fstab = image_rootfs + "/etc/fstab" if not os.path.isfile(fstab): return None parts = self._get_parts() self._save_fstab(fstab) fstab_lines = self._get_fstab(fstab, parts) self._update_fstab(fstab_lines, parts) self._write_fstab(fstab, fstab_lines) return fstab def _update_fstab(self, fstab_lines, parts): """Assume partition order same as in wks""" for num, p in enumerate(parts, 1): if not p.mountpoint or p.mountpoint == "/" or p.mountpoint == "/boot": continue if self._ptable_format == 'msdos' and num > 3: device_name = "/dev/" + p.disk + str(num + 1) else: device_name = "/dev/" + p.disk + str(num) fstab_entry = device_name + "\t" + p.mountpoint + "\t" + p.fstype + "\tdefaults\t0\t0\n" fstab_lines.append(fstab_entry) def _write_fstab(self, fstab, fstab_lines): fstab = open(fstab, "w") for line in fstab_lines: fstab.write(line) fstab.close() def _save_fstab(self, fstab): """Save the current fstab in rootfs""" shutil.copyfile(fstab, fstab + ".orig") def _restore_fstab(self, fstab): """Restore the saved fstab in rootfs""" if fstab is None: return shutil.move(fstab + ".orig", fstab) def _get_fstab(self, fstab, parts): """Return the desired contents of /etc/fstab.""" f = open(fstab, "r") fstab_contents = f.readlines() f.close() return fstab_contents def set_bootimg_dir(self, bootimg_dir): """ Accessor for bootimg_dir, the actual location used for the source of the bootimg. Should be set by source plugins (only if they change the default bootimg source) so the correct info gets displayed for print_outimage_info(). """ self.bootimg_dir = bootimg_dir def _get_parts(self): if not self.ks: raise CreatorError("Failed to get partition info, " "please check your kickstart setting.") # Set a default partition if no partition is given out if not self.ks.handler.partition.partitions: partstr = "part / --size 1900 --ondisk sda --fstype=ext3" args = partstr.split() pd = self.ks.handler.partition.parse(args[1:]) if pd not in self.ks.handler.partition.partitions: self.ks.handler.partition.partitions.append(pd) # partitions list from kickstart file return kickstart.get_partitions(self.ks) def get_disk_names(self): """ Returns a list of physical target disk names (e.g., 'sdb') which will be created. """ if self._disk_names: return self._disk_names #get partition info from ks handler parts = self._get_parts() for i in range(len(parts)): if parts[i].disk: disk_name = parts[i].disk else: raise CreatorError("Failed to create disks, no --ondisk " "specified in partition line of ks file") if parts[i].mountpoint and not parts[i].fstype: raise CreatorError("Failed to create disks, no --fstype " "specified for partition with mountpoint " "'%s' in the ks file") self._disk_names.append(disk_name) return self._disk_names def _full_name(self, name, extention): """ Construct full file name for a file we generate. """ return "%s-%s.%s" % (self.name, name, extention) def _full_path(self, path, name, extention): """ Construct full file path to a file we generate. """ return os.path.join(path, self._full_name(name, extention)) def get_default_source_plugin(self): """ The default source plugin i.e. the plugin that's consulted for overall image generation tasks outside of any particular partition. For convenience, we just hang it off the bootloader handler since it's the one non-partition object in any setup. By default the default plugin is set to the same plugin as the /boot partition; since we hang it off the bootloader object, the default can be explicitly set using the --source bootloader param. """ return self.ks.handler.bootloader.source # # Actual implemention # def _mount_instroot(self, base_on = None): """ For 'wic', we already have our build artifacts and don't want to loop mount anything to install into, we just create filesystems from the artifacts directly and combine them into a partitioned image. We still want to reuse as much of the basic mic machinery though; despite the fact that we don't actually do loop or any other kind of mounting we still want to do many of the same things to prepare images, so we basically just adapt to the basic framework and reinterpret what 'mounting' means in our context. _instroot would normally be something like /var/tmp/wic/build/imgcreate-s_9AKQ/install_root, for installing packages, etc. We don't currently need to do that, so we simplify life by just using /var/tmp/wic/build as our workdir. """ parts = self._get_parts() self.__instimage = PartitionedMount(self._instroot) for p in parts: # as a convenience, set source to the boot partition source # instead of forcing it to be set via bootloader --source if not self.ks.handler.bootloader.source and p.mountpoint == "/boot": self.ks.handler.bootloader.source = p.source for p in parts: # need to create the filesystems in order to get their # sizes before we can add them and do the layout. # PartitionedMount.mount() actually calls __format_disks() # to create the disk images and carve out the partitions, # then self.install() calls PartitionedMount.install() # which calls __install_partitition() for each partition # to dd the fs into the partitions. It would be nice to # be able to use e.g. ExtDiskMount etc to create the # filesystems, since that's where existing e.g. mkfs code # is, but those are only created after __format_disks() # which needs the partition sizes so needs them created # before its called. Well, the existing setup is geared # to installing packages into mounted filesystems - maybe # when/if we need to actually do package selection we # should modify things to use those objects, but for now # we can avoid that. p.prepare(self, self.workdir, self.oe_builddir, self.rootfs_dir, self.bootimg_dir, self.kernel_dir, self.native_sysroot) fstab = self.__write_fstab(p.get_rootfs()) self._restore_fstab(fstab) self.__instimage.add_partition(int(p.size), p.disk, p.mountpoint, p.source_file, p.fstype, p.label, fsopts = p.fsopts, boot = p.active, align = p.align, part_type = p.part_type) self.__instimage.layout_partitions(self._ptable_format) self.__imgdir = self.workdir for disk_name, disk in self.__instimage.disks.items(): full_path = self._full_path(self.__imgdir, disk_name, "direct") msger.debug("Adding disk %s as %s with size %s bytes" \ % (disk_name, full_path, disk['min_size'])) disk_obj = fs_related.DiskImage(full_path, disk['min_size']) self.__disks[disk_name] = disk_obj self.__instimage.add_disk(disk_name, disk_obj) self.__instimage.mount() def install(self, repo_urls=None): """ Install fs images into partitions """ for disk_name, disk in self.__instimage.disks.items(): full_path = self._full_path(self.__imgdir, disk_name, "direct") msger.debug("Installing disk %s as %s with size %s bytes" \ % (disk_name, full_path, disk['min_size'])) self.__instimage.install(full_path) def configure(self, repodata = None): """ Configure the system image according to kickstart. For now, it just prepares the image to be bootable by e.g. creating and installing a bootloader configuration. """ source_plugin = self.get_default_source_plugin() if source_plugin: self._source_methods = pluginmgr.get_source_plugin_methods(source_plugin, disk_methods) for disk_name, disk in self.__instimage.disks.items(): self._source_methods["do_install_disk"](disk, disk_name, self, self.workdir, self.oe_builddir, self.bootimg_dir, self.kernel_dir, self.native_sysroot) def print_outimage_info(self): """ Print the image(s) and artifacts used, for the user. """ msg = "The new image(s) can be found here:\n" parts = self._get_parts() for disk_name, disk in self.__instimage.disks.items(): full_path = self._full_path(self.__imgdir, disk_name, "direct") msg += ' %s\n\n' % full_path msg += 'The following build artifacts were used to create the image(s):\n' for p in parts: if p.get_rootfs() is None: continue if p.mountpoint == '/': str = ':' else: str = '["%s"]:' % p.label msg += ' ROOTFS_DIR%s%s\n' % (str.ljust(20), p.get_rootfs()) msg += ' BOOTIMG_DIR: %s\n' % self.bootimg_dir msg += ' KERNEL_DIR: %s\n' % self.kernel_dir msg += ' NATIVE_SYSROOT: %s\n' % self.native_sysroot msger.info(msg) def _get_boot_config(self): """ Return the rootdev/root_part_uuid (if specified by --part-type) Assume partition order same as in wks """ rootdev = None root_part_uuid = None parts = self._get_parts() for num, p in enumerate(parts, 1): if p.mountpoint == "/": part = '' if p.disk.startswith('mmcblk'): part = 'p' if self._ptable_format == 'msdos' and num > 3: rootdev = "/dev/%s%s%-d" % (p.disk, part, num + 1) else: rootdev = "/dev/%s%s%-d" % (p.disk, part, num) root_part_uuid = p.part_type return (rootdev, root_part_uuid) def _unmount_instroot(self): if not self.__instimage is None: try: self.__instimage.cleanup() except MountError, err: msger.warning("%s" % err) ########NEW FILE######## __FILENAME__ = fs #!/usr/bin/python -tt # # Copyright (c) 2011 Intel, Inc. # # This program is free software; you can redistribute it and/or modify it # under the terms of the GNU General Public License as published by the Free # Software Foundation; version 2 of the License # # This program is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY # or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License # for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., 59 # Temple Place - Suite 330, Boston, MA 02111-1307, USA. import os from mic import msger from mic.utils import runner, misc from mic.utils.errors import CreatorError from mic.utils.fs_related import find_binary_path from mic.imager.baseimager import BaseImageCreator class FsImageCreator(BaseImageCreator): def __init__(self, cfgmgr = None, pkgmgr = None): self.zips = { "tar.bz2" : "" } BaseImageCreator.__init__(self, cfgmgr, pkgmgr) self._fstype = None self._fsopts = None self._include_src = False def package(self, destdir = "."): ignores = ["/dev/fd", "/dev/stdin", "/dev/stdout", "/dev/stderr", "/etc/mtab"] if not os.path.exists(destdir): os.makedirs(destdir) if self._recording_pkgs: self._save_recording_pkgs(destdir) if not self.pack_to: fsdir = os.path.join(destdir, self.name) misc.check_space_pre_cp(self._instroot, destdir) msger.info("Copying %s to %s ..." % (self._instroot, fsdir)) runner.show(['cp', "-af", self._instroot, fsdir]) for exclude in ignores: if os.path.exists(fsdir + exclude): os.unlink(fsdir + exclude) self.outimage.append(fsdir) else: (tar, comp) = os.path.splitext(self.pack_to) try: tarcreat = {'.tar': '-cf', '.gz': '-czf', '.bz2': '-cjf', '.tgz': '-czf', '.tbz': '-cjf'}[comp] except KeyError: raise CreatorError("Unsupported comression for this image type:" " '%s', try '.tar', '.tar.gz', etc" % comp) dst = os.path.join(destdir, self.pack_to) msger.info("Pack rootfs to %s. Please wait..." % dst) tar = find_binary_path('tar') tar_cmdline = [tar, "--numeric-owner", "--preserve-permissions", "--preserve-order", "--one-file-system", "--directory", self._instroot] for ignore_entry in ignores: if ignore_entry.startswith('/'): ignore_entry = ignore_entry[1:] tar_cmdline.append("--exclude=%s" % (ignore_entry)) tar_cmdline.extend([tarcreat, dst, "."]) rc = runner.show(tar_cmdline) if rc: raise CreatorError("Failed compress image with tar.bz2. " "Cmdline: %s" % (" ".join(tar_cmdline))) self.outimage.append(dst) ########NEW FILE######## __FILENAME__ = livecd #!/usr/bin/python -tt # # Copyright (c) 2011 Intel, Inc. # # This program is free software; you can redistribute it and/or modify it # under the terms of the GNU General Public License as published by the Free # Software Foundation; version 2 of the License # # This program is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY # or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License # for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., 59 # Temple Place - Suite 330, Boston, MA 02111-1307, USA. import os, sys import glob import shutil from mic import kickstart, msger from mic.utils import fs_related, runner, misc from mic.utils.errors import CreatorError from mic.imager.loop import LoopImageCreator class LiveImageCreatorBase(LoopImageCreator): """A base class for LiveCD image creators. This class serves as a base class for the architecture-specific LiveCD image creator subclass, LiveImageCreator. LiveImageCreator creates a bootable ISO containing the system image, bootloader, bootloader configuration, kernel and initramfs. """ def __init__(self, creatoropts = None, pkgmgr = None): """Initialise a LiveImageCreator instance. This method takes the same arguments as ImageCreator.__init__(). """ LoopImageCreator.__init__(self, creatoropts, pkgmgr) #Controls whether to use squashfs to compress the image. self.skip_compression = False #Controls whether an image minimizing snapshot should be created. # #This snapshot can be used when copying the system image from the ISO in #order to minimize the amount of data that needs to be copied; simply, #it makes it possible to create a version of the image's filesystem with #no spare space. self.skip_minimize = False #A flag which indicates i act as a convertor default false self.actasconvertor = False #The bootloader timeout from kickstart. if self.ks: self._timeout = kickstart.get_timeout(self.ks, 10) else: self._timeout = 10 #The default kernel type from kickstart. if self.ks: self._default_kernel = kickstart.get_default_kernel(self.ks, "kernel") else: self._default_kernel = None if self.ks: parts = kickstart.get_partitions(self.ks) if len(parts) > 1: raise CreatorError("Can't support multi partitions in ks file " "for this image type") # FIXME: rename rootfs img to self.name, # else can't find files when create iso self._instloops[0]['name'] = self.name + ".img" self.__isodir = None self.__modules = ["=ata", "sym53c8xx", "aic7xxx", "=usb", "=firewire", "=mmc", "=pcmcia", "mptsas"] if self.ks: self.__modules.extend(kickstart.get_modules(self.ks)) self._dep_checks.extend(["isohybrid", "unsquashfs", "mksquashfs", "dd", "genisoimage"]) # # Hooks for subclasses # def _configure_bootloader(self, isodir): """Create the architecture specific booloader configuration. This is the hook where subclasses must create the booloader configuration in order to allow a bootable ISO to be built. isodir -- the directory where the contents of the ISO are to be staged """ raise CreatorError("Bootloader configuration is arch-specific, " "but not implemented for this arch!") def _get_menu_options(self): """Return a menu options string for syslinux configuration. """ if self.ks is None: return "liveinst autoinst" r = kickstart.get_menu_args(self.ks) return r def _get_kernel_options(self): """Return a kernel options string for bootloader configuration. This is the hook where subclasses may specify a set of kernel options which should be included in the images bootloader configuration. A sensible default implementation is provided. """ if self.ks is None: r = "ro rd.live.image" else: r = kickstart.get_kernel_args(self.ks) return r def _get_mkisofs_options(self, isodir): """Return the architecture specific mkisosfs options. This is the hook where subclasses may specify additional arguments to mkisofs, e.g. to enable a bootable ISO to be built. By default, an empty list is returned. """ return [] # # Helpers for subclasses # def _has_checkisomd5(self): """Check whether checkisomd5 is available in the install root.""" def _exists(path): return os.path.exists(self._instroot + path) if _exists("/usr/bin/checkisomd5") and os.path.exists("/usr/bin/implantisomd5"): return True return False def __restore_file(self,path): try: os.unlink(path) except: pass if os.path.exists(path + '.rpmnew'): os.rename(path + '.rpmnew', path) def _mount_instroot(self, base_on = None): LoopImageCreator._mount_instroot(self, base_on) self.__write_initrd_conf(self._instroot + "/etc/sysconfig/mkinitrd") self.__write_dracut_conf(self._instroot + "/etc/dracut.conf.d/02livecd.conf") def _unmount_instroot(self): self.__restore_file(self._instroot + "/etc/sysconfig/mkinitrd") self.__restore_file(self._instroot + "/etc/dracut.conf.d/02livecd.conf") LoopImageCreator._unmount_instroot(self) def __ensure_isodir(self): if self.__isodir is None: self.__isodir = self._mkdtemp("iso-") return self.__isodir def _get_isodir(self): return self.__ensure_isodir() def _set_isodir(self, isodir = None): self.__isodir = isodir def _create_bootconfig(self): """Configure the image so that it's bootable.""" self._configure_bootloader(self.__ensure_isodir()) def _get_post_scripts_env(self, in_chroot): env = LoopImageCreator._get_post_scripts_env(self, in_chroot) if not in_chroot: env["LIVE_ROOT"] = self.__ensure_isodir() return env def __write_dracut_conf(self, path): if not os.path.exists(os.path.dirname(path)): fs_related.makedirs(os.path.dirname(path)) f = open(path, "a") f.write('add_dracutmodules+=" dmsquash-live pollcdrom "') f.close() def __write_initrd_conf(self, path): content = "" if not os.path.exists(os.path.dirname(path)): fs_related.makedirs(os.path.dirname(path)) f = open(path, "w") content += 'LIVEOS="yes"\n' content += 'PROBE="no"\n' content += 'MODULES+="squashfs ext3 ext2 vfat msdos "\n' content += 'MODULES+="sr_mod sd_mod ide-cd cdrom "\n' for module in self.__modules: if module == "=usb": content += 'MODULES+="ehci_hcd uhci_hcd ohci_hcd "\n' content += 'MODULES+="usb_storage usbhid "\n' elif module == "=firewire": content += 'MODULES+="firewire-sbp2 firewire-ohci "\n' content += 'MODULES+="sbp2 ohci1394 ieee1394 "\n' elif module == "=mmc": content += 'MODULES+="mmc_block sdhci sdhci-pci "\n' elif module == "=pcmcia": content += 'MODULES+="pata_pcmcia "\n' else: content += 'MODULES+="' + module + ' "\n' f.write(content) f.close() def __create_iso(self, isodir): iso = self._outdir + "/" + self.name + ".iso" genisoimage = fs_related.find_binary_path("genisoimage") args = [genisoimage, "-J", "-r", "-hide-rr-moved", "-hide-joliet-trans-tbl", "-V", self.fslabel, "-o", iso] args.extend(self._get_mkisofs_options(isodir)) args.append(isodir) if runner.show(args) != 0: raise CreatorError("ISO creation failed!") """ It should be ok still even if you haven't isohybrid """ isohybrid = None try: isohybrid = fs_related.find_binary_path("isohybrid") except: pass if isohybrid: args = [isohybrid, "-partok", iso ] if runner.show(args) != 0: raise CreatorError("Hybrid ISO creation failed!") self.__implant_md5sum(iso) def __implant_md5sum(self, iso): """Implant an isomd5sum.""" if os.path.exists("/usr/bin/implantisomd5"): implantisomd5 = "/usr/bin/implantisomd5" else: msger.warning("isomd5sum not installed; not setting up mediacheck") implantisomd5 = "" return runner.show([implantisomd5, iso]) def _stage_final_image(self): try: fs_related.makedirs(self.__ensure_isodir() + "/LiveOS") minimal_size = self._resparse() if not self.skip_minimize: fs_related.create_image_minimizer(self.__isodir + \ "/LiveOS/osmin.img", self._image, minimal_size) if self.skip_compression: shutil.move(self._image, self.__isodir + "/LiveOS/ext3fs.img") else: fs_related.makedirs(os.path.join( os.path.dirname(self._image), "LiveOS")) shutil.move(self._image, os.path.join(os.path.dirname(self._image), "LiveOS", "ext3fs.img")) fs_related.mksquashfs(os.path.dirname(self._image), self.__isodir + "/LiveOS/squashfs.img") self.__create_iso(self.__isodir) if self.pack_to: isoimg = os.path.join(self._outdir, self.name + ".iso") packimg = os.path.join(self._outdir, self.pack_to) misc.packing(packimg, isoimg) os.unlink(isoimg) finally: shutil.rmtree(self.__isodir, ignore_errors = True) self.__isodir = None class x86LiveImageCreator(LiveImageCreatorBase): """ImageCreator for x86 machines""" def _get_mkisofs_options(self, isodir): return [ "-b", "isolinux/isolinux.bin", "-c", "isolinux/boot.cat", "-no-emul-boot", "-boot-info-table", "-boot-load-size", "4" ] def _get_required_packages(self): return ["syslinux", "syslinux-extlinux"] + \ LiveImageCreatorBase._get_required_packages(self) def _get_isolinux_stanzas(self, isodir): return "" def __find_syslinux_menu(self): for menu in ["vesamenu.c32", "menu.c32"]: if os.path.isfile(self._instroot + "/usr/share/syslinux/" + menu): return menu raise CreatorError("syslinux not installed : " "no suitable /usr/share/syslinux/*menu.c32 found") def __find_syslinux_mboot(self): # # We only need the mboot module if we have any xen hypervisors # if not glob.glob(self._instroot + "/boot/xen.gz*"): return None return "mboot.c32" def __copy_syslinux_files(self, isodir, menu, mboot = None): files = ["isolinux.bin", menu] if mboot: files += [mboot] for f in files: path = self._instroot + "/usr/share/syslinux/" + f if not os.path.isfile(path): raise CreatorError("syslinux not installed : " "%s not found" % path) shutil.copy(path, isodir + "/isolinux/") def __copy_syslinux_background(self, isodest): background_path = self._instroot + \ "/usr/share/branding/default/syslinux/syslinux-vesa-splash.jpg" if not os.path.exists(background_path): return False shutil.copyfile(background_path, isodest) return True def __copy_kernel_and_initramfs(self, isodir, version, index): bootdir = self._instroot + "/boot" isDracut = False if self._alt_initrd_name: src_initrd_path = os.path.join(bootdir, self._alt_initrd_name) else: if os.path.exists(bootdir + "/initramfs-" + version + ".img"): src_initrd_path = os.path.join(bootdir, "initramfs-" +version+ ".img") isDracut = True else: src_initrd_path = os.path.join(bootdir, "initrd-" +version+ ".img") try: msger.debug("copy %s to %s" % (bootdir + "/vmlinuz-" + version, isodir + "/isolinux/vmlinuz" + index)) shutil.copyfile(bootdir + "/vmlinuz-" + version, isodir + "/isolinux/vmlinuz" + index) msger.debug("copy %s to %s" % (src_initrd_path, isodir + "/isolinux/initrd" + index + ".img")) shutil.copyfile(src_initrd_path, isodir + "/isolinux/initrd" + index + ".img") except: raise CreatorError("Unable to copy valid kernels or initrds, " "please check the repo.") is_xen = False if os.path.exists(bootdir + "/xen.gz-" + version[:-3]): shutil.copyfile(bootdir + "/xen.gz-" + version[:-3], isodir + "/isolinux/xen" + index + ".gz") is_xen = True return (is_xen,isDracut) def __is_default_kernel(self, kernel, kernels): if len(kernels) == 1: return True if kernel == self._default_kernel: return True if kernel.startswith("kernel-") and kernel[7:] == self._default_kernel: return True return False def __get_basic_syslinux_config(self, **args): return """ default %(menu)s timeout %(timeout)d %(background)s menu title Welcome to %(distroname)s! menu color border 0 #ffffffff #00000000 menu color sel 7 #ff000000 #ffffffff menu color title 0 #ffffffff #00000000 menu color tabmsg 0 #ffffffff #00000000 menu color unsel 0 #ffffffff #00000000 menu color hotsel 0 #ff000000 #ffffffff menu color hotkey 7 #ffffffff #ff000000 menu color timeout_msg 0 #ffffffff #00000000 menu color timeout 0 #ffffffff #00000000 menu color cmdline 0 #ffffffff #00000000 menu hidden menu clear """ % args def __get_image_stanza(self, is_xen, isDracut, **args): if isDracut: args["rootlabel"] = "live:CDLABEL=%(fslabel)s" % args else: args["rootlabel"] = "CDLABEL=%(fslabel)s" % args if not is_xen: template = """label %(short)s menu label %(long)s kernel vmlinuz%(index)s append initrd=initrd%(index)s.img root=%(rootlabel)s rootfstype=iso9660 %(liveargs)s %(extra)s """ else: template = """label %(short)s menu label %(long)s kernel mboot.c32 append xen%(index)s.gz --- vmlinuz%(index)s root=%(rootlabel)s rootfstype=iso9660 %(liveargs)s %(extra)s --- initrd%(index)s.img """ return template % args def __get_image_stanzas(self, isodir): versions = [] kernels = self._get_kernel_versions() for kernel in kernels: for version in kernels[kernel]: versions.append(version) if not versions: raise CreatorError("Unable to find valid kernels, " "please check the repo") kernel_options = self._get_kernel_options() """ menu can be customized highly, the format is: short_name1:long_name1:extra_opts1;short_name2:long_name2:extra_opts2 e.g.: autoinst:InstallationOnly:systemd.unit=installer-graphical.service but in order to keep compatible with old format, these are still ok: liveinst autoinst liveinst;autoinst liveinst::;autoinst:: """ oldmenus = {"basic": { "short": "basic", "long": "Installation Only (Text based)", "extra": "basic nosplash 4" }, "liveinst": { "short": "liveinst", "long": "Installation Only", "extra": "liveinst nosplash 4" }, "autoinst": { "short": "autoinst", "long": "Autoinstall (Deletes all existing content)", "extra": "autoinst nosplash 4" }, "netinst": { "short": "netinst", "long": "Network Installation", "extra": "netinst 4" }, "verify": { "short": "check", "long": "Verify and", "extra": "check" } } menu_options = self._get_menu_options() menus = menu_options.split(";") for i in range(len(menus)): menus[i] = menus[i].split(":") if len(menus) == 1 and len(menus[0]) == 1: """ Keep compatible with the old usage way """ menus = menu_options.split() for i in range(len(menus)): menus[i] = [menus[i]] cfg = "" default_version = None default_index = None index = "0" netinst = None for version in versions: (is_xen, isDracut) = self.__copy_kernel_and_initramfs(isodir, version, index) if index == "0": self._isDracut = isDracut default = self.__is_default_kernel(kernel, kernels) if default: long = "Boot %s" % self.distro_name elif kernel.startswith("kernel-"): long = "Boot %s(%s)" % (self.name, kernel[7:]) else: long = "Boot %s(%s)" % (self.name, kernel) oldmenus["verify"]["long"] = "%s %s" % (oldmenus["verify"]["long"], long) # tell dracut not to ask for LUKS passwords or activate mdraid sets if isDracut: kern_opts = kernel_options + " rd.luks=0 rd.md=0 rd.dm=0" else: kern_opts = kernel_options cfg += self.__get_image_stanza(is_xen, isDracut, fslabel = self.fslabel, liveargs = kern_opts, long = long, short = "linux" + index, extra = "", index = index) if default: cfg += "menu default\n" default_version = version default_index = index for menu in menus: if not menu[0]: continue short = menu[0] + index if len(menu) >= 2: long = menu[1] else: if menu[0] in oldmenus.keys(): if menu[0] == "verify" and not self._has_checkisomd5(): continue if menu[0] == "netinst": netinst = oldmenus[menu[0]] continue long = oldmenus[menu[0]]["long"] extra = oldmenus[menu[0]]["extra"] else: long = short.upper() + " X" + index extra = "" if len(menu) >= 3: extra = menu[2] cfg += self.__get_image_stanza(is_xen, isDracut, fslabel = self.fslabel, liveargs = kernel_options, long = long, short = short, extra = extra, index = index) index = str(int(index) + 1) if not default_version: default_version = versions[0] if not default_index: default_index = "0" if netinst: cfg += self.__get_image_stanza(is_xen, isDracut, fslabel = self.fslabel, liveargs = kernel_options, long = netinst["long"], short = netinst["short"], extra = netinst["extra"], index = default_index) return cfg def __get_memtest_stanza(self, isodir): memtest = glob.glob(self._instroot + "/boot/memtest86*") if not memtest: return "" shutil.copyfile(memtest[0], isodir + "/isolinux/memtest") return """label memtest menu label Memory Test kernel memtest """ def __get_local_stanza(self, isodir): return """label local menu label Boot from local drive localboot 0xffff """ def _configure_syslinux_bootloader(self, isodir): """configure the boot loader""" fs_related.makedirs(isodir + "/isolinux") menu = self.__find_syslinux_menu() self.__copy_syslinux_files(isodir, menu, self.__find_syslinux_mboot()) background = "" if self.__copy_syslinux_background(isodir + "/isolinux/splash.jpg"): background = "menu background splash.jpg" cfg = self.__get_basic_syslinux_config(menu = menu, background = background, name = self.name, timeout = self._timeout * 10, distroname = self.distro_name) cfg += self.__get_image_stanzas(isodir) cfg += self.__get_memtest_stanza(isodir) cfg += self.__get_local_stanza(isodir) cfg += self._get_isolinux_stanzas(isodir) cfgf = open(isodir + "/isolinux/isolinux.cfg", "w") cfgf.write(cfg) cfgf.close() def __copy_efi_files(self, isodir): if not os.path.exists(self._instroot + "/boot/efi/EFI/redhat/grub.efi"): return False shutil.copy(self._instroot + "/boot/efi/EFI/redhat/grub.efi", isodir + "/EFI/boot/grub.efi") shutil.copy(self._instroot + "/boot/grub/splash.xpm.gz", isodir + "/EFI/boot/splash.xpm.gz") return True def __get_basic_efi_config(self, **args): return """ default=0 splashimage=/EFI/boot/splash.xpm.gz timeout %(timeout)d hiddenmenu """ %args def __get_efi_image_stanza(self, **args): return """title %(long)s kernel /EFI/boot/vmlinuz%(index)s root=CDLABEL=%(fslabel)s rootfstype=iso9660 %(liveargs)s %(extra)s initrd /EFI/boot/initrd%(index)s.img """ %args def __get_efi_image_stanzas(self, isodir, name): # FIXME: this only supports one kernel right now... kernel_options = self._get_kernel_options() checkisomd5 = self._has_checkisomd5() cfg = "" for index in range(0, 9): # we don't support xen kernels if os.path.exists("%s/EFI/boot/xen%d.gz" %(isodir, index)): continue cfg += self.__get_efi_image_stanza(fslabel = self.fslabel, liveargs = kernel_options, long = name, extra = "", index = index) if checkisomd5: cfg += self.__get_efi_image_stanza( fslabel = self.fslabel, liveargs = kernel_options, long = "Verify and Boot " + name, extra = "check", index = index) break return cfg def _configure_efi_bootloader(self, isodir): """Set up the configuration for an EFI bootloader""" fs_related.makedirs(isodir + "/EFI/boot") if not self.__copy_efi_files(isodir): shutil.rmtree(isodir + "/EFI") return for f in os.listdir(isodir + "/isolinux"): os.link("%s/isolinux/%s" %(isodir, f), "%s/EFI/boot/%s" %(isodir, f)) cfg = self.__get_basic_efi_config(name = self.name, timeout = self._timeout) cfg += self.__get_efi_image_stanzas(isodir, self.name) cfgf = open(isodir + "/EFI/boot/grub.conf", "w") cfgf.write(cfg) cfgf.close() # first gen mactel machines get the bootloader name wrong apparently if rpmmisc.getBaseArch() == "i386": os.link(isodir + "/EFI/boot/grub.efi", isodir + "/EFI/boot/boot.efi") os.link(isodir + "/EFI/boot/grub.conf", isodir + "/EFI/boot/boot.conf") # for most things, we want them named boot$efiarch efiarch = {"i386": "ia32", "x86_64": "x64"} efiname = efiarch[rpmmisc.getBaseArch()] os.rename(isodir + "/EFI/boot/grub.efi", isodir + "/EFI/boot/boot%s.efi" %(efiname,)) os.link(isodir + "/EFI/boot/grub.conf", isodir + "/EFI/boot/boot%s.conf" %(efiname,)) def _configure_bootloader(self, isodir): self._configure_syslinux_bootloader(isodir) self._configure_efi_bootloader(isodir) arch = "i386" if arch in ("i386", "x86_64"): LiveCDImageCreator = x86LiveImageCreator elif arch.startswith("arm"): LiveCDImageCreator = LiveImageCreatorBase else: raise CreatorError("Architecture not supported!") ########NEW FILE######## __FILENAME__ = liveusb #!/usr/bin/python -tt # # Copyright (c) 2011 Intel, Inc. # # This program is free software; you can redistribute it and/or modify it # under the terms of the GNU General Public License as published by the Free # Software Foundation; version 2 of the License # # This program is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY # or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License # for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., 59 # Temple Place - Suite 330, Boston, MA 02111-1307, USA. import os import shutil import re from mic import msger from mic.utils import misc, fs_related, runner from mic.utils.errors import CreatorError, MountError from mic.utils.partitionedfs import PartitionedMount from mic.imager.livecd import LiveCDImageCreator class LiveUSBImageCreator(LiveCDImageCreator): def __init__(self, *args): LiveCDImageCreator.__init__(self, *args) self._dep_checks.extend(["kpartx", "parted"]) # remove dependency of genisoimage in parent class if "genisoimage" in self._dep_checks: self._dep_checks.remove("genisoimage") def _create_usbimg(self, isodir): overlaysizemb = 64 #default #skipcompress = self.skip_compression? fstype = "vfat" homesizemb=0 swapsizemb=0 homefile="home.img" plussize=128 kernelargs=None if fstype == 'vfat': if overlaysizemb > 2047: raise CreatorError("Can't have an overlay of 2048MB or " "greater on VFAT") if homesizemb > 2047: raise CreatorError("Can't have an home overlay of 2048MB or " "greater on VFAT") if swapsizemb > 2047: raise CreatorError("Can't have an swap overlay of 2048MB or " "greater on VFAT") livesize = misc.get_file_size(isodir + "/LiveOS") usbimgsize = (overlaysizemb + \ homesizemb + \ swapsizemb + \ livesize + \ plussize) * 1024L * 1024L disk = fs_related.SparseLoopbackDisk("%s/%s.usbimg" \ % (self._outdir, self.name), usbimgsize) usbmnt = self._mkdtemp("usb-mnt") usbloop = PartitionedMount(usbmnt) usbloop.add_disk('/dev/sdb', disk) usbloop.add_partition(usbimgsize/1024/1024, "/dev/sdb", "/", fstype, boot=True) usbloop.mount() try: fs_related.makedirs(usbmnt + "/LiveOS") if os.path.exists(isodir + "/LiveOS/squashfs.img"): shutil.copyfile(isodir + "/LiveOS/squashfs.img", usbmnt + "/LiveOS/squashfs.img") else: fs_related.mksquashfs(os.path.dirname(self._image), usbmnt + "/LiveOS/squashfs.img") if os.path.exists(isodir + "/LiveOS/osmin.img"): shutil.copyfile(isodir + "/LiveOS/osmin.img", usbmnt + "/LiveOS/osmin.img") if fstype == "vfat" or fstype == "msdos": uuid = usbloop.partitions[0]['mount'].uuid label = usbloop.partitions[0]['mount'].fslabel usblabel = "UUID=%s-%s" % (uuid[0:4], uuid[4:8]) overlaysuffix = "-%s-%s-%s" % (label, uuid[0:4], uuid[4:8]) else: diskmount = usbloop.partitions[0]['mount'] usblabel = "UUID=%s" % diskmount.uuid overlaysuffix = "-%s-%s" % (diskmount.fslabel, diskmount.uuid) args = ['cp', "-Rf", isodir + "/isolinux", usbmnt + "/syslinux"] rc = runner.show(args) if rc: raise CreatorError("Can't copy isolinux directory %s" \ % (isodir + "/isolinux/*")) if os.path.isfile("/usr/share/syslinux/isolinux.bin"): syslinux_path = "/usr/share/syslinux" elif os.path.isfile("/usr/lib/syslinux/isolinux.bin"): syslinux_path = "/usr/lib/syslinux" else: raise CreatorError("syslinux not installed : " "cannot find syslinux installation path") for f in ("isolinux.bin", "vesamenu.c32"): path = os.path.join(syslinux_path, f) if os.path.isfile(path): args = ['cp', path, usbmnt + "/syslinux/"] rc = runner.show(args) if rc: raise CreatorError("Can't copy syslinux file " + path) else: raise CreatorError("syslinux not installed: " "syslinux file %s not found" % path) fd = open(isodir + "/isolinux/isolinux.cfg", "r") text = fd.read() fd.close() pattern = re.compile('CDLABEL=[^ ]*') text = pattern.sub(usblabel, text) pattern = re.compile('rootfstype=[^ ]*') text = pattern.sub("rootfstype=" + fstype, text) if kernelargs: text = text.replace("rd.live.image", "rd.live.image " + kernelargs) if overlaysizemb > 0: msger.info("Initializing persistent overlay file") overfile = "overlay" + overlaysuffix if fstype == "vfat": args = ['dd', "if=/dev/zero", "of=" + usbmnt + "/LiveOS/" + overfile, "count=%d" % overlaysizemb, "bs=1M"] else: args = ['dd', "if=/dev/null", "of=" + usbmnt + "/LiveOS/" + overfile, "count=1", "bs=1M", "seek=%d" % overlaysizemb] rc = runner.show(args) if rc: raise CreatorError("Can't create overlay file") text = text.replace("rd.live.image", "rd.live.image rd.live.overlay=" + usblabel) text = text.replace(" ro ", " rw ") if swapsizemb > 0: msger.info("Initializing swap file") swapfile = usbmnt + "/LiveOS/" + "swap.img" args = ['dd', "if=/dev/zero", "of=" + swapfile, "count=%d" % swapsizemb, "bs=1M"] rc = runner.show(args) if rc: raise CreatorError("Can't create swap file") args = ["mkswap", "-f", swapfile] rc = runner.show(args) if rc: raise CreatorError("Can't mkswap on swap file") if homesizemb > 0: msger.info("Initializing persistent /home") homefile = usbmnt + "/LiveOS/" + homefile if fstype == "vfat": args = ['dd', "if=/dev/zero", "of=" + homefile, "count=%d" % homesizemb, "bs=1M"] else: args = ['dd', "if=/dev/null", "of=" + homefile, "count=1", "bs=1M", "seek=%d" % homesizemb] rc = runner.show(args) if rc: raise CreatorError("Can't create home file") mkfscmd = fs_related.find_binary_path("/sbin/mkfs." + fstype) if fstype == "ext2" or fstype == "ext3": args = [mkfscmd, "-F", "-j", homefile] else: args = [mkfscmd, homefile] rc = runner.show(args) if rc: raise CreatorError("Can't mke2fs home file") if fstype == "ext2" or fstype == "ext3": tune2fs = fs_related.find_binary_path("tune2fs") args = [tune2fs, "-c0", "-i0", "-ouser_xattr,acl", homefile] rc = runner.show(args) if rc: raise CreatorError("Can't tune2fs home file") if fstype == "vfat" or fstype == "msdos": syslinuxcmd = fs_related.find_binary_path("syslinux") syslinuxcfg = usbmnt + "/syslinux/syslinux.cfg" args = [syslinuxcmd, "-d", "syslinux", usbloop.partitions[0]["device"]] elif fstype == "ext2" or fstype == "ext3": extlinuxcmd = fs_related.find_binary_path("extlinux") syslinuxcfg = usbmnt + "/syslinux/extlinux.conf" args = [extlinuxcmd, "-i", usbmnt + "/syslinux"] else: raise CreatorError("Invalid file system type: %s" % (fstype)) os.unlink(usbmnt + "/syslinux/isolinux.cfg") fd = open(syslinuxcfg, "w") fd.write(text) fd.close() rc = runner.show(args) if rc: raise CreatorError("Can't install boot loader.") finally: usbloop.unmount() usbloop.cleanup() # Need to do this after image is unmounted and device mapper is closed msger.info("set MBR") mbrfile = "/usr/lib/syslinux/mbr.bin" if not os.path.exists(mbrfile): mbrfile = "/usr/share/syslinux/mbr.bin" if not os.path.exists(mbrfile): raise CreatorError("mbr.bin file didn't exist.") mbrsize = os.path.getsize(mbrfile) outimg = "%s/%s.usbimg" % (self._outdir, self.name) args = ['dd', "if=" + mbrfile, "of=" + outimg, "seek=0", "conv=notrunc", "bs=1", "count=%d" % (mbrsize)] rc = runner.show(args) if rc: raise CreatorError("Can't set MBR.") def _stage_final_image(self): try: isodir = self._get_isodir() fs_related.makedirs(isodir + "/LiveOS") minimal_size = self._resparse() if not self.skip_minimize: fs_related.create_image_minimizer(isodir + "/LiveOS/osmin.img", self._image, minimal_size) if self.skip_compression: shutil.move(self._image, isodir + "/LiveOS/ext3fs.img") else: fs_related.makedirs(os.path.join( os.path.dirname(self._image), "LiveOS")) shutil.move(self._image, os.path.join(os.path.dirname(self._image), "LiveOS", "ext3fs.img")) fs_related.mksquashfs(os.path.dirname(self._image), isodir + "/LiveOS/squashfs.img") self._create_usbimg(isodir) if self.pack_to: usbimg = os.path.join(self._outdir, self.name + ".usbimg") packimg = os.path.join(self._outdir, self.pack_to) misc.packing(packimg, usbimg) os.unlink(usbimg) finally: shutil.rmtree(isodir, ignore_errors = True) self._set_isodir(None) ########NEW FILE######## __FILENAME__ = loop #!/usr/bin/python -tt # # Copyright (c) 2011 Intel, Inc. # # This program is free software; you can redistribute it and/or modify it # under the terms of the GNU General Public License as published by the Free # Software Foundation; version 2 of the License # # This program is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY # or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License # for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., 59 # Temple Place - Suite 330, Boston, MA 02111-1307, USA. import os import glob import shutil from mic import kickstart, msger from mic.utils.errors import CreatorError, MountError from mic.utils import misc, runner, fs_related as fs from mic.imager.baseimager import BaseImageCreator # The maximum string length supported for LoopImageCreator.fslabel FSLABEL_MAXLEN = 32 def save_mountpoints(fpath, loops, arch = None): """Save mount points mapping to file :fpath, the xml file to store partition info :loops, dict of partition info :arch, image arch """ if not fpath or not loops: return from xml.dom import minidom doc = minidom.Document() imgroot = doc.createElement("image") doc.appendChild(imgroot) if arch: imgroot.setAttribute('arch', arch) for loop in loops: part = doc.createElement("partition") imgroot.appendChild(part) for (key, val) in loop.items(): if isinstance(val, fs.Mount): continue part.setAttribute(key, str(val)) with open(fpath, 'w') as wf: wf.write(doc.toprettyxml(indent=' ')) return def load_mountpoints(fpath): """Load mount points mapping from file :fpath, file path to load """ if not fpath: return from xml.dom import minidom mount_maps = [] with open(fpath, 'r') as rf: dom = minidom.parse(rf) imgroot = dom.documentElement for part in imgroot.getElementsByTagName("partition"): p = dict(part.attributes.items()) try: mp = (p['mountpoint'], p['label'], p['name'], int(p['size']), p['fstype']) except KeyError: msger.warning("Wrong format line in file: %s" % fpath) except ValueError: msger.warning("Invalid size '%s' in file: %s" % (p['size'], fpath)) else: mount_maps.append(mp) return mount_maps class LoopImageCreator(BaseImageCreator): """Installs a system into a loopback-mountable filesystem image. LoopImageCreator is a straightforward ImageCreator subclass; the system is installed into an ext3 filesystem on a sparse file which can be subsequently loopback-mounted. When specifying multiple partitions in kickstart file, each partition will be created as a separated loop image. """ def __init__(self, creatoropts=None, pkgmgr=None, compress_image=None, shrink_image=False): """Initialize a LoopImageCreator instance. This method takes the same arguments as ImageCreator.__init__() with the addition of: fslabel -- A string used as a label for any filesystems created. """ BaseImageCreator.__init__(self, creatoropts, pkgmgr) self.compress_image = compress_image self.shrink_image = shrink_image self.__fslabel = None self.fslabel = self.name self.__blocksize = 4096 if self.ks: self.__fstype = kickstart.get_image_fstype(self.ks, "ext3") self.__fsopts = kickstart.get_image_fsopts(self.ks, "defaults,noatime") allloops = [] for part in sorted(kickstart.get_partitions(self.ks), key=lambda p: p.mountpoint): if part.fstype == "swap": continue label = part.label mp = part.mountpoint if mp == '/': # the base image if not label: label = self.name else: mp = mp.rstrip('/') if not label: msger.warning('no "label" specified for loop img at %s' ', use the mountpoint as the name' % mp) label = mp.split('/')[-1] imgname = misc.strip_end(label, '.img') + '.img' allloops.append({ 'mountpoint': mp, 'label': label, 'name': imgname, 'size': part.size or 4096L * 1024 * 1024, 'fstype': part.fstype or 'ext3', 'extopts': part.extopts or None, 'loop': None, # to be created in _mount_instroot }) self._instloops = allloops else: self.__fstype = None self.__fsopts = None self._instloops = [] self.__imgdir = None if self.ks: self.__image_size = kickstart.get_image_size(self.ks, 4096L * 1024 * 1024) else: self.__image_size = 0 self._img_name = self.name + ".img" def get_image_names(self): if not self._instloops: return None return [lo['name'] for lo in self._instloops] def _set_fstype(self, fstype): self.__fstype = fstype def _set_image_size(self, imgsize): self.__image_size = imgsize # # Properties # def __get_fslabel(self): if self.__fslabel is None: return self.name else: return self.__fslabel def __set_fslabel(self, val): if val is None: self.__fslabel = None else: self.__fslabel = val[:FSLABEL_MAXLEN] #A string used to label any filesystems created. # #Some filesystems impose a constraint on the maximum allowed size of the #filesystem label. In the case of ext3 it's 16 characters, but in the case #of ISO9660 it's 32 characters. # #mke2fs silently truncates the label, but mkisofs aborts if the label is #too long. So, for convenience sake, any string assigned to this attribute #is silently truncated to FSLABEL_MAXLEN (32) characters. fslabel = property(__get_fslabel, __set_fslabel) def __get_image(self): if self.__imgdir is None: raise CreatorError("_image is not valid before calling mount()") return os.path.join(self.__imgdir, self._img_name) #The location of the image file. # #This is the path to the filesystem image. Subclasses may use this path #in order to package the image in _stage_final_image(). # #Note, this directory does not exist before ImageCreator.mount() is called. # #Note also, this is a read-only attribute. _image = property(__get_image) def __get_blocksize(self): return self.__blocksize def __set_blocksize(self, val): if self._instloops: raise CreatorError("_blocksize must be set before calling mount()") try: self.__blocksize = int(val) except ValueError: raise CreatorError("'%s' is not a valid integer value " "for _blocksize" % val) #The block size used by the image's filesystem. # #This is the block size used when creating the filesystem image. Subclasses #may change this if they wish to use something other than a 4k block size. # #Note, this attribute may only be set before calling mount(). _blocksize = property(__get_blocksize, __set_blocksize) def __get_fstype(self): return self.__fstype def __set_fstype(self, val): if val != "ext2" and val != "ext3": raise CreatorError("Unknown _fstype '%s' supplied" % val) self.__fstype = val #The type of filesystem used for the image. # #This is the filesystem type used when creating the filesystem image. #Subclasses may change this if they wish to use something other ext3. # #Note, only ext2 and ext3 are currently supported. # #Note also, this attribute may only be set before calling mount(). _fstype = property(__get_fstype, __set_fstype) def __get_fsopts(self): return self.__fsopts def __set_fsopts(self, val): self.__fsopts = val #Mount options of filesystem used for the image. # #This can be specified by --fsoptions="xxx,yyy" in part command in #kickstart file. _fsopts = property(__get_fsopts, __set_fsopts) # # Helpers for subclasses # def _resparse(self, size=None): """Rebuild the filesystem image to be as sparse as possible. This method should be used by subclasses when staging the final image in order to reduce the actual space taken up by the sparse image file to be as little as possible. This is done by resizing the filesystem to the minimal size (thereby eliminating any space taken up by deleted files) and then resizing it back to the supplied size. size -- the size in, in bytes, which the filesystem image should be resized to after it has been minimized; this defaults to None, causing the original size specified by the kickstart file to be used (or 4GiB if not specified in the kickstart). """ minsize = 0 for item in self._instloops: if item['name'] == self._img_name: minsize = item['loop'].resparse(size) else: item['loop'].resparse(size) return minsize def _base_on(self, base_on=None): if base_on and self._image != base_on: shutil.copyfile(base_on, self._image) def _check_imgdir(self): if self.__imgdir is None: self.__imgdir = self._mkdtemp() # # Actual implementation # def _mount_instroot(self, base_on=None): if base_on and os.path.isfile(base_on): self.__imgdir = os.path.dirname(base_on) imgname = os.path.basename(base_on) self._base_on(base_on) self._set_image_size(misc.get_file_size(self._image)) # here, self._instloops must be [] self._instloops.append({ "mountpoint": "/", "label": self.name, "name": imgname, "size": self.__image_size or 4096L, "fstype": self.__fstype or "ext3", "extopts": None, "loop": None }) self._check_imgdir() for loop in self._instloops: fstype = loop['fstype'] mp = os.path.join(self._instroot, loop['mountpoint'].lstrip('/')) size = loop['size'] * 1024L * 1024L imgname = loop['name'] if fstype in ("ext2", "ext3", "ext4"): MyDiskMount = fs.ExtDiskMount elif fstype == "btrfs": MyDiskMount = fs.BtrfsDiskMount elif fstype in ("vfat", "msdos"): MyDiskMount = fs.VfatDiskMount else: msger.error('Cannot support fstype: %s' % fstype) loop['loop'] = MyDiskMount(fs.SparseLoopbackDisk( os.path.join(self.__imgdir, imgname), size), mp, fstype, self._blocksize, loop['label']) if fstype in ("ext2", "ext3", "ext4"): loop['loop'].extopts = loop['extopts'] try: msger.verbose('Mounting image "%s" on "%s"' % (imgname, mp)) fs.makedirs(mp) loop['loop'].mount() except MountError, e: raise def _unmount_instroot(self): for item in reversed(self._instloops): try: item['loop'].cleanup() except: pass def _stage_final_image(self): if self.pack_to or self.shrink_image: self._resparse(0) else: self._resparse() for item in self._instloops: imgfile = os.path.join(self.__imgdir, item['name']) if item['fstype'] == "ext4": runner.show('/sbin/tune2fs -O ^huge_file,extents,uninit_bg %s ' % imgfile) if self.compress_image: misc.compressing(imgfile, self.compress_image) if not self.pack_to: for item in os.listdir(self.__imgdir): shutil.move(os.path.join(self.__imgdir, item), os.path.join(self._outdir, item)) else: msger.info("Pack all loop images together to %s" % self.pack_to) dstfile = os.path.join(self._outdir, self.pack_to) misc.packing(dstfile, self.__imgdir) if self.pack_to: mountfp_xml = os.path.splitext(self.pack_to)[0] mountfp_xml = misc.strip_end(mountfp_xml, '.tar') + ".xml" else: mountfp_xml = self.name + ".xml" # save mount points mapping file to xml save_mountpoints(os.path.join(self._outdir, mountfp_xml), self._instloops, self.target_arch) def copy_attachment(self): if not hasattr(self, '_attachment') or not self._attachment: return self._check_imgdir() msger.info("Copying attachment files...") for item in self._attachment: if not os.path.exists(item): continue dpath = os.path.join(self.__imgdir, os.path.basename(item)) msger.verbose("Copy attachment %s to %s" % (item, dpath)) shutil.copy(item, dpath) ########NEW FILE######## __FILENAME__ = raw #!/usr/bin/python -tt # # Copyright (c) 2011 Intel, Inc. # # This program is free software; you can redistribute it and/or modify it # under the terms of the GNU General Public License as published by the Free # Software Foundation; version 2 of the License # # This program is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY # or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License # for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., 59 # Temple Place - Suite 330, Boston, MA 02111-1307, USA. import os import stat import shutil from mic import kickstart, msger from mic.utils import fs_related, runner, misc from mic.utils.partitionedfs import PartitionedMount from mic.utils.errors import CreatorError, MountError from mic.imager.baseimager import BaseImageCreator class RawImageCreator(BaseImageCreator): """Installs a system into a file containing a partitioned disk image. ApplianceImageCreator is an advanced ImageCreator subclass; a sparse file is formatted with a partition table, each partition loopback mounted and the system installed into an virtual disk. The disk image can subsequently be booted in a virtual machine or accessed with kpartx """ def __init__(self, creatoropts=None, pkgmgr=None, compress_image=None, generate_bmap=None, fstab_entry="uuid"): """Initialize a ApplianceImageCreator instance. This method takes the same arguments as ImageCreator.__init__() """ BaseImageCreator.__init__(self, creatoropts, pkgmgr) self.__instloop = None self.__imgdir = None self.__disks = {} self.__disk_format = "raw" self._disk_names = [] self._ptable_format = self.ks.handler.bootloader.ptable self.vmem = 512 self.vcpu = 1 self.checksum = False self.use_uuid = fstab_entry == "uuid" self.appliance_version = None self.appliance_release = None self.compress_image = compress_image self.bmap_needed = generate_bmap self._need_extlinux = not kickstart.use_installerfw(self.ks, "extlinux") #self.getsource = False #self.listpkg = False self._dep_checks.extend(["sync", "kpartx", "parted"]) if self._need_extlinux: self._dep_checks.extend(["extlinux"]) def configure(self, repodata = None): import subprocess def chroot(): os.chroot(self._instroot) os.chdir("/") if os.path.exists(self._instroot + "/usr/bin/Xorg"): subprocess.call(["/bin/chmod", "u+s", "/usr/bin/Xorg"], preexec_fn = chroot) BaseImageCreator.configure(self, repodata) def _get_fstab(self): if kickstart.use_installerfw(self.ks, "fstab"): # The fstab file will be generated by installer framework scripts # instead. return None s = "" for mp in self.__instloop.mountOrder: p = None for p1 in self.__instloop.partitions: if p1['mountpoint'] == mp: p = p1 break if self.use_uuid and p['uuid']: device = "UUID=%s" % p['uuid'] else: device = "/dev/%s%-d" % (p['disk_name'], p['num']) s += "%(device)s %(mountpoint)s %(fstype)s %(fsopts)s 0 0\n" % { 'device': device, 'mountpoint': p['mountpoint'], 'fstype': p['fstype'], 'fsopts': "defaults,noatime" if not p['fsopts'] else p['fsopts']} if p['mountpoint'] == "/": for subvol in self.__instloop.subvolumes: if subvol['mountpoint'] == "/": continue s += "%(device)s %(mountpoint)s %(fstype)s %(fsopts)s 0 0\n" % { 'device': "/dev/%s%-d" % (p['disk_name'], p['num']), 'mountpoint': subvol['mountpoint'], 'fstype': p['fstype'], 'fsopts': "defaults,noatime" if not subvol['fsopts'] else subvol['fsopts']} s += "devpts /dev/pts devpts gid=5,mode=620 0 0\n" s += "tmpfs /dev/shm tmpfs defaults 0 0\n" s += "proc /proc proc defaults 0 0\n" s += "sysfs /sys sysfs defaults 0 0\n" return s def _create_mkinitrd_config(self): """write to tell which modules to be included in initrd""" mkinitrd = "" mkinitrd += "PROBE=\"no\"\n" mkinitrd += "MODULES+=\"ext3 ata_piix sd_mod libata scsi_mod\"\n" mkinitrd += "rootfs=\"ext3\"\n" mkinitrd += "rootopts=\"defaults\"\n" msger.debug("Writing mkinitrd config %s/etc/sysconfig/mkinitrd" \ % self._instroot) os.makedirs(self._instroot + "/etc/sysconfig/",mode=644) cfg = open(self._instroot + "/etc/sysconfig/mkinitrd", "w") cfg.write(mkinitrd) cfg.close() def _get_parts(self): if not self.ks: raise CreatorError("Failed to get partition info, " "please check your kickstart setting.") # Set a default partition if no partition is given out if not self.ks.handler.partition.partitions: partstr = "part / --size 1900 --ondisk sda --fstype=ext3" args = partstr.split() pd = self.ks.handler.partition.parse(args[1:]) if pd not in self.ks.handler.partition.partitions: self.ks.handler.partition.partitions.append(pd) # partitions list from kickstart file return kickstart.get_partitions(self.ks) def get_disk_names(self): """ Returns a list of physical target disk names (e.g., 'sdb') which will be created. """ if self._disk_names: return self._disk_names #get partition info from ks handler parts = self._get_parts() for i in range(len(parts)): if parts[i].disk: disk_name = parts[i].disk else: raise CreatorError("Failed to create disks, no --ondisk " "specified in partition line of ks file") if parts[i].mountpoint and not parts[i].fstype: raise CreatorError("Failed to create disks, no --fstype " "specified for partition with mountpoint " "'%s' in the ks file") self._disk_names.append(disk_name) return self._disk_names def _full_name(self, name, extention): """ Construct full file name for a file we generate. """ return "%s-%s.%s" % (self.name, name, extention) def _full_path(self, path, name, extention): """ Construct full file path to a file we generate. """ return os.path.join(path, self._full_name(name, extention)) # # Actual implemention # def _mount_instroot(self, base_on = None): parts = self._get_parts() self.__instloop = PartitionedMount(self._instroot) for p in parts: self.__instloop.add_partition(int(p.size), p.disk, p.mountpoint, p.fstype, p.label, fsopts = p.fsopts, boot = p.active, align = p.align, part_type = p.part_type) self.__instloop.layout_partitions(self._ptable_format) # Create the disks self.__imgdir = self._mkdtemp() for disk_name, disk in self.__instloop.disks.items(): full_path = self._full_path(self.__imgdir, disk_name, "raw") msger.debug("Adding disk %s as %s with size %s bytes" \ % (disk_name, full_path, disk['min_size'])) disk_obj = fs_related.SparseLoopbackDisk(full_path, disk['min_size']) self.__disks[disk_name] = disk_obj self.__instloop.add_disk(disk_name, disk_obj) self.__instloop.mount() self._create_mkinitrd_config() def _get_required_packages(self): required_packages = BaseImageCreator._get_required_packages(self) if self._need_extlinux: if not self.target_arch or not self.target_arch.startswith("arm"): required_packages += ["syslinux", "syslinux-extlinux"] return required_packages def _get_excluded_packages(self): return BaseImageCreator._get_excluded_packages(self) def _get_syslinux_boot_config(self): rootdev = None root_part_uuid = None for p in self.__instloop.partitions: if p['mountpoint'] == "/": rootdev = "/dev/%s%-d" % (p['disk_name'], p['num']) root_part_uuid = p['partuuid'] return (rootdev, root_part_uuid) def _create_syslinux_config(self): splash = os.path.join(self._instroot, "boot/extlinux") if os.path.exists(splash): splashline = "menu background splash.jpg" else: splashline = "" (rootdev, root_part_uuid) = self._get_syslinux_boot_config() options = self.ks.handler.bootloader.appendLine #XXX don't hardcode default kernel - see livecd code syslinux_conf = "" syslinux_conf += "prompt 0\n" syslinux_conf += "timeout 1\n" syslinux_conf += "\n" syslinux_conf += "default vesamenu.c32\n" syslinux_conf += "menu autoboot Starting %s...\n" % self.distro_name syslinux_conf += "menu hidden\n" syslinux_conf += "\n" syslinux_conf += "%s\n" % splashline syslinux_conf += "menu title Welcome to %s!\n" % self.distro_name syslinux_conf += "menu color border 0 #ffffffff #00000000\n" syslinux_conf += "menu color sel 7 #ffffffff #ff000000\n" syslinux_conf += "menu color title 0 #ffffffff #00000000\n" syslinux_conf += "menu color tabmsg 0 #ffffffff #00000000\n" syslinux_conf += "menu color unsel 0 #ffffffff #00000000\n" syslinux_conf += "menu color hotsel 0 #ff000000 #ffffffff\n" syslinux_conf += "menu color hotkey 7 #ffffffff #ff000000\n" syslinux_conf += "menu color timeout_msg 0 #ffffffff #00000000\n" syslinux_conf += "menu color timeout 0 #ffffffff #00000000\n" syslinux_conf += "menu color cmdline 0 #ffffffff #00000000\n" versions = [] kernels = self._get_kernel_versions() symkern = "%s/boot/vmlinuz" % self._instroot if os.path.lexists(symkern): v = os.path.realpath(symkern).replace('%s-' % symkern, "") syslinux_conf += "label %s\n" % self.distro_name.lower() syslinux_conf += "\tmenu label %s (%s)\n" % (self.distro_name, v) syslinux_conf += "\tlinux ../vmlinuz\n" if self._ptable_format == 'msdos': rootstr = rootdev else: if not root_part_uuid: raise MountError("Cannot find the root GPT partition UUID") rootstr = "PARTUUID=%s" % root_part_uuid syslinux_conf += "\tappend ro root=%s %s\n" % (rootstr, options) syslinux_conf += "\tmenu default\n" else: for kernel in kernels: for version in kernels[kernel]: versions.append(version) footlabel = 0 for v in versions: syslinux_conf += "label %s%d\n" \ % (self.distro_name.lower(), footlabel) syslinux_conf += "\tmenu label %s (%s)\n" % (self.distro_name, v) syslinux_conf += "\tlinux ../vmlinuz-%s\n" % v syslinux_conf += "\tappend ro root=%s %s\n" \ % (rootdev, options) if footlabel == 0: syslinux_conf += "\tmenu default\n" footlabel += 1; msger.debug("Writing syslinux config %s/boot/extlinux/extlinux.conf" \ % self._instroot) cfg = open(self._instroot + "/boot/extlinux/extlinux.conf", "w") cfg.write(syslinux_conf) cfg.close() def _install_syslinux(self): for name in self.__disks.keys(): loopdev = self.__disks[name].device # Set MBR mbrfile = "%s/usr/share/syslinux/" % self._instroot if self._ptable_format == 'gpt': mbrfile += "gptmbr.bin" else: mbrfile += "mbr.bin" msger.debug("Installing syslinux bootloader '%s' to %s" % \ (mbrfile, loopdev)) mbrsize = os.stat(mbrfile)[stat.ST_SIZE] rc = runner.show(['dd', 'if=%s' % mbrfile, 'of=' + loopdev]) if rc != 0: raise MountError("Unable to set MBR to %s" % loopdev) # Ensure all data is flushed to disk before doing syslinux install runner.quiet('sync') fullpathsyslinux = fs_related.find_binary_path("extlinux") rc = runner.show([fullpathsyslinux, "-i", "%s/boot/extlinux" % self._instroot]) if rc != 0: raise MountError("Unable to install syslinux bootloader to %s" \ % loopdev) def _create_bootconfig(self): #If syslinux is available do the required configurations. if self._need_extlinux \ and os.path.exists("%s/usr/share/syslinux/" % (self._instroot)) \ and os.path.exists("%s/boot/extlinux/" % (self._instroot)): self._create_syslinux_config() self._install_syslinux() def _unmount_instroot(self): if not self.__instloop is None: try: self.__instloop.cleanup() except MountError, err: msger.warning("%s" % err) def _resparse(self, size = None): return self.__instloop.resparse(size) def _get_post_scripts_env(self, in_chroot): env = BaseImageCreator._get_post_scripts_env(self, in_chroot) # Export the file-system UUIDs and partition UUIDs (AKA PARTUUIDs) for p in self.__instloop.partitions: env.update(self._set_part_env(p['ks_pnum'], "UUID", p['uuid'])) env.update(self._set_part_env(p['ks_pnum'], "PARTUUID", p['partuuid'])) return env def _stage_final_image(self): """Stage the final system image in _outdir. write meta data """ self._resparse() if self.compress_image: for imgfile in os.listdir(self.__imgdir): if imgfile.endswith('.raw') or imgfile.endswith('bin'): imgpath = os.path.join(self.__imgdir, imgfile) misc.compressing(imgpath, self.compress_image) if self.pack_to: dst = os.path.join(self._outdir, self.pack_to) msger.info("Pack all raw images to %s" % dst) misc.packing(dst, self.__imgdir) else: msger.debug("moving disks to stage location") for imgfile in os.listdir(self.__imgdir): src = os.path.join(self.__imgdir, imgfile) dst = os.path.join(self._outdir, imgfile) msger.debug("moving %s to %s" % (src,dst)) shutil.move(src,dst) self._write_image_xml() def _write_image_xml(self): imgarch = "i686" if self.target_arch and self.target_arch.startswith("arm"): imgarch = "arm" xml = "<image>\n" name_attributes = "" if self.appliance_version: name_attributes += " version='%s'" % self.appliance_version if self.appliance_release: name_attributes += " release='%s'" % self.appliance_release xml += " <name%s>%s</name>\n" % (name_attributes, self.name) xml += " <domain>\n" # XXX don't hardcode - determine based on the kernel we installed for # grub baremetal vs xen xml += " <boot type='hvm'>\n" xml += " <guest>\n" xml += " <arch>%s</arch>\n" % imgarch xml += " </guest>\n" xml += " <os>\n" xml += " <loader dev='hd'/>\n" xml += " </os>\n" i = 0 for name in self.__disks.keys(): full_name = self._full_name(name, self.__disk_format) xml += " <drive disk='%s' target='hd%s'/>\n" \ % (full_name, chr(ord('a') + i)) i = i + 1 xml += " </boot>\n" xml += " <devices>\n" xml += " <vcpu>%s</vcpu>\n" % self.vcpu xml += " <memory>%d</memory>\n" %(self.vmem * 1024) for network in self.ks.handler.network.network: xml += " <interface/>\n" xml += " <graphics/>\n" xml += " </devices>\n" xml += " </domain>\n" xml += " <storage>\n" if self.checksum is True: for name in self.__disks.keys(): diskpath = self._full_path(self._outdir, name, \ self.__disk_format) full_name = self._full_name(name, self.__disk_format) msger.debug("Generating disk signature for %s" % full_name) xml += " <disk file='%s' use='system' format='%s'>\n" \ % (full_name, self.__disk_format) hashes = misc.calc_hashes(diskpath, ('sha1', 'sha256')) xml += " <checksum type='sha1'>%s</checksum>\n" \ % hashes[0] xml += " <checksum type='sha256'>%s</checksum>\n" \ % hashes[1] xml += " </disk>\n" else: for name in self.__disks.keys(): full_name = self._full_name(name, self.__disk_format) xml += " <disk file='%s' use='system' format='%s'/>\n" \ % (full_name, self.__disk_format) xml += " </storage>\n" xml += "</image>\n" msger.debug("writing image XML to %s/%s.xml" %(self._outdir, self.name)) cfg = open("%s/%s.xml" % (self._outdir, self.name), "w") cfg.write(xml) cfg.close() def generate_bmap(self): """ Generate block map file for the image. The idea is that while disk images we generate may be large (e.g., 4GiB), they may actually contain only little real data, e.g., 512MiB. This data are files, directories, file-system meta-data, partition table, etc. In other words, when flashing the image to the target device, you do not have to copy all the 4GiB of data, you can copy only 512MiB of it, which is 4 times faster. This function generates the block map file for an arbitrary image that mic has generated. The block map file is basically an XML file which contains a list of blocks which have to be copied to the target device. The other blocks are not used and there is no need to copy them. """ if self.bmap_needed is None: return from mic.utils import BmapCreate msger.info("Generating the map file(s)") for name in self.__disks.keys(): image = self._full_path(self.__imgdir, name, self.__disk_format) bmap_file = self._full_path(self._outdir, name, "bmap") msger.debug("Generating block map file '%s'" % bmap_file) try: creator = BmapCreate.BmapCreate(image, bmap_file) creator.generate() del creator except BmapCreate.Error as err: raise CreatorError("Failed to create bmap file: %s" % str(err)) ########NEW FILE######## __FILENAME__ = desktop #!/usr/bin/python -tt # # Copyright (c) 2008, 2009, 2010 Intel, Inc. # # Yi Yang <yi.y.yang@intel.com> # # This program is free software; you can redistribute it and/or modify it # under the terms of the GNU General Public License as published by the Free # Software Foundation; version 2 of the License # # This program is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY # or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License # for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., 59 # Temple Place - Suite 330, Boston, MA 02111-1307, USA. from pykickstart.base import * from pykickstart.errors import * from pykickstart.options import * class Mic_Desktop(KickstartCommand): def __init__(self, writePriority=0, defaultdesktop=None, defaultdm=None, autologinuser=None, session=None): KickstartCommand.__init__(self, writePriority) self.__new_version = False self.op = self._getParser() self.defaultdesktop = defaultdesktop self.autologinuser = autologinuser self.defaultdm = defaultdm self.session = session def __str__(self): retval = "" if self.defaultdesktop != None: retval += " --defaultdesktop=%s" % self.defaultdesktop if self.session != None: retval += " --session=\"%s\"" % self.session if self.autologinuser != None: retval += " --autologinuser=%s" % self.autologinuser if self.defaultdm != None: retval += " --defaultdm=%s" % self.defaultdm if retval != "": retval = "# Default Desktop Settings\ndesktop %s\n" % retval return retval def _getParser(self): try: op = KSOptionParser(lineno=self.lineno) except TypeError: # the latest version has not lineno argument op = KSOptionParser() self.__new_version = True op.add_option("--defaultdesktop", dest="defaultdesktop", action="store", type="string", nargs=1) op.add_option("--autologinuser", dest="autologinuser", action="store", type="string", nargs=1) op.add_option("--defaultdm", dest="defaultdm", action="store", type="string", nargs=1) op.add_option("--session", dest="session", action="store", type="string", nargs=1) return op def parse(self, args): if self.__new_version: (opts, extra) = self.op.parse_args(args=args, lineno=self.lineno) else: (opts, extra) = self.op.parse_args(args=args) if extra: m = _("Unexpected arguments to %(command)s command: %(options)s") \ % {"command": "desktop", "options": extra} raise KickstartValueError, formatErrorMsg(self.lineno, msg=m) self._setToSelf(self.op, opts) ########NEW FILE######## __FILENAME__ = installerfw #!/usr/bin/python -tt # # Copyright (c) 2013 Intel, Inc. # # This program is free software; you can redistribute it and/or modify it # under the terms of the GNU General Public License as published by the Free # Software Foundation; version 2 of the License # # This program is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY # or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License # for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., 59 # Temple Place - Suite 330, Boston, MA 02111-1307, USA. from pykickstart.base import * from pykickstart.options import * class Mic_installerfw(KickstartCommand): """ This class implements the "installerfw" KS option. The argument of the option is a comman-separated list of MIC features which have to be disabled and instead, will be done in the installer. For example, "installerfw=extlinux" disables all the MIC code which installs extlinux to the target images, and instead, the extlinux or whatever boot-loader will be installed by the installer instead. The installer is a tool which is external to MIC, it comes from the installation repositories and can be executed by MIC in order to perform various configuration actions. The main point here is to make sure MIC has no hard-wired knoledge about the target OS configuration. """ removedKeywords = KickstartCommand.removedKeywords removedAttrs = KickstartCommand.removedAttrs def __init__(self, *args, **kwargs): KickstartCommand.__init__(self, *args, **kwargs) self.op = self._getParser() self.features = kwargs.get("installerfw", None) def __str__(self): retval = KickstartCommand.__str__(self) if self.features: retval += "# Enable installer framework features\ninstallerfw\n" return retval def _getParser(self): op = KSOptionParser() return op def parse(self, args): (_, extra) = self.op.parse_args(args=args, lineno=self.lineno) if len(extra) != 1: msg = "Kickstart command \"installerfw\" requires one " \ "argumet - a list of legacy features to disable" raise KickstartValueError, formatErrorMsg(self.lineno, msg = msg) self.features = extra[0].split(",") return self ########NEW FILE######## __FILENAME__ = micboot #!/usr/bin/python -tt # # Copyright (c) 2008, 2009, 2010 Intel, Inc. # # Anas Nashif # # This program is free software; you can redistribute it and/or modify it # under the terms of the GNU General Public License as published by the Free # Software Foundation; version 2 of the License # # This program is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY # or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License # for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., 59 # Temple Place - Suite 330, Boston, MA 02111-1307, USA. from pykickstart.base import * from pykickstart.errors import * from pykickstart.options import * from pykickstart.commands.bootloader import * class Mic_Bootloader(F8_Bootloader): def __init__(self, writePriority=10, appendLine="", driveorder=None, forceLBA=False, location="", md5pass="", password="", upgrade=False, menus=""): F8_Bootloader.__init__(self, writePriority, appendLine, driveorder, forceLBA, location, md5pass, password, upgrade) self.menus = "" self.ptable = "msdos" def _getArgsAsStr(self): ret = F8_Bootloader._getArgsAsStr(self) if self.menus == "": ret += " --menus=%s" %(self.menus,) if self.ptable: ret += " --ptable=\"%s\"" %(self.ptable,) return ret def _getParser(self): op = F8_Bootloader._getParser(self) op.add_option("--menus", dest="menus") op.add_option("--ptable", dest="ptable", type="string") return op ########NEW FILE######## __FILENAME__ = micpartition #!/usr/bin/python -tt # # Marko Saukko <marko.saukko@cybercom.com> # # Copyright (C) 2011 Nokia Corporation and/or its subsidiary(-ies). # # This copyrighted material is made available to anyone wishing to use, modify, # copy, or redistribute it subject to the terms and conditions of the GNU # General Public License v.2. This program is distributed in the hope that it # will be useful, but WITHOUT ANY WARRANTY expressed or implied, including the # implied warranties of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., 51 # Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. from pykickstart.commands.partition import * class Mic_PartData(FC4_PartData): removedKeywords = FC4_PartData.removedKeywords removedAttrs = FC4_PartData.removedAttrs def __init__(self, *args, **kwargs): FC4_PartData.__init__(self, *args, **kwargs) self.deleteRemovedAttrs() self.align = kwargs.get("align", None) self.extopts = kwargs.get("extopts", None) self.part_type = kwargs.get("part_type", None) def _getArgsAsStr(self): retval = FC4_PartData._getArgsAsStr(self) if self.align: retval += " --align" if self.extopts: retval += " --extoptions=%s" % self.extopts if self.part_type: retval += " --part-type=%s" % self.part_type return retval class Mic_Partition(FC4_Partition): removedKeywords = FC4_Partition.removedKeywords removedAttrs = FC4_Partition.removedAttrs def _getParser(self): op = FC4_Partition._getParser(self) # The alignment value is given in kBytes. e.g., value 8 means that # the partition is aligned to start from 8096 byte boundary. op.add_option("--align", type="int", action="store", dest="align", default=None) op.add_option("--extoptions", type="string", action="store", dest="extopts", default=None) op.add_option("--part-type", type="string", action="store", dest="part_type", default=None) return op ########NEW FILE######## __FILENAME__ = micrepo #!/usr/bin/python -tt # # Copyright (c) 2008, 2009, 2010 Intel, Inc. # # Yi Yang <yi.y.yang@intel.com> # # This program is free software; you can redistribute it and/or modify it # under the terms of the GNU General Public License as published by the Free # Software Foundation; version 2 of the License # # This program is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY # or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License # for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., 59 # Temple Place - Suite 330, Boston, MA 02111-1307, USA. from pykickstart.base import * from pykickstart.errors import * from pykickstart.options import * from pykickstart.commands.repo import * class Mic_RepoData(F8_RepoData): def __init__(self, baseurl="", mirrorlist=None, name="", priority=None, includepkgs=(), excludepkgs=(), save=False, proxy=None, proxy_username=None, proxy_password=None, debuginfo=False, source=False, gpgkey=None, disable=False, ssl_verify="yes", nocache=False): kw = {} # F8_RepoData keywords if includepkgs: kw['includepkgs'] = includepkgs if excludepkgs: kw['excludepkgs'] = excludepkgs #FC6_RepoData keywords if baseurl: kw['baseurl'] = baseurl if mirrorlist: kw['mirrorlist'] = mirrorlist if name: kw['name'] = name F8_RepoData.__init__(self, **kw) self.save = save self.proxy = proxy self.proxy_username = proxy_username self.proxy_password = proxy_password self.debuginfo = debuginfo self.disable = disable self.source = source self.gpgkey = gpgkey self.ssl_verify = ssl_verify.lower() self.priority = priority self.nocache = nocache def _getArgsAsStr(self): retval = F8_RepoData._getArgsAsStr(self) if self.save: retval += " --save" if self.proxy: retval += " --proxy=%s" % self.proxy if self.proxy_username: retval += " --proxyuser=%s" % self.proxy_username if self.proxy_password: retval += " --proxypasswd=%s" % self.proxy_password if self.debuginfo: retval += " --debuginfo" if self.source: retval += " --source" if self.gpgkey: retval += " --gpgkey=%s" % self.gpgkey if self.disable: retval += " --disable" if self.ssl_verify: retval += " --ssl_verify=%s" % self.ssl_verify if self.priority: retval += " --priority=%s" % self.priority if self.nocache: retval += " --nocache" return retval class Mic_Repo(F8_Repo): def __init__(self, writePriority=0, repoList=None): F8_Repo.__init__(self, writePriority, repoList) def __str__(self): retval = "" for repo in self.repoList: retval += repo.__str__() return retval def _getParser(self): def list_cb (option, opt_str, value, parser): for d in value.split(','): parser.values.ensure_value(option.dest, []).append(d) op = F8_Repo._getParser(self) op.add_option("--save", action="store_true", dest="save", default=False) op.add_option("--proxy", type="string", action="store", dest="proxy", default=None, nargs=1) op.add_option("--proxyuser", type="string", action="store", dest="proxy_username", default=None, nargs=1) op.add_option("--proxypasswd", type="string", action="store", dest="proxy_password", default=None, nargs=1) op.add_option("--debuginfo", action="store_true", dest="debuginfo", default=False) op.add_option("--source", action="store_true", dest="source", default=False) op.add_option("--disable", action="store_true", dest="disable", default=False) op.add_option("--gpgkey", type="string", action="store", dest="gpgkey", default=None, nargs=1) op.add_option("--ssl_verify", type="string", action="store", dest="ssl_verify", default="yes") op.add_option("--priority", type="int", action="store", dest="priority", default=None) op.add_option("--nocache", action="store_true", dest="nocache", default=False) return op ########NEW FILE######## __FILENAME__ = partition # ex:ts=4:sw=4:sts=4:et # -*- tab-width: 4; c-basic-offset: 4; indent-tabs-mode: nil -*- # # Copyright (c) 2013, Intel Corporation. # All rights reserved. # # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License version 2 as # published by the Free Software Foundation. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., # 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA. # # DESCRIPTION # This module provides the OpenEmbedded partition object definitions. # # AUTHORS # Tom Zanussi <tom.zanussi (at] linux.intel.com> # import shutil from pykickstart.commands.partition import * from mic.utils.oe.misc import * from mic.kickstart.custom_commands import * from mic.plugin import pluginmgr partition_methods = { "do_stage_partition":None, "do_prepare_partition":None, "do_configure_partition":None, } class Wic_PartData(Mic_PartData): removedKeywords = Mic_PartData.removedKeywords removedAttrs = Mic_PartData.removedAttrs def __init__(self, *args, **kwargs): Mic_PartData.__init__(self, *args, **kwargs) self.deleteRemovedAttrs() self.source = kwargs.get("source", None) self.rootfs = kwargs.get("rootfs-dir", None) self.source_file = "" self.size = 0 def _getArgsAsStr(self): retval = Mic_PartData._getArgsAsStr(self) if self.source: retval += " --source=%s" % self.source if self.rootfs: retval += " --rootfs-dir=%s" % self.rootfs return retval def get_rootfs(self): """ Acessor for rootfs dir """ return self.rootfs def set_rootfs(self, rootfs): """ Acessor for actual rootfs dir, which must be set by source plugins. """ self.rootfs = rootfs def get_size(self): """ Accessor for partition size, 0 or --size before set_size(). """ return self.size def set_size(self, size): """ Accessor for actual partition size, which must be set by source plugins. """ self.size = size def set_source_file(self, source_file): """ Accessor for source_file, the location of the generated partition image, which must be set by source plugins. """ self.source_file = source_file def get_extra_block_count(self, current_blocks): """ The --size param is reflected in self.size (in MB), and we already have current_blocks (1k) blocks, calculate and return the number of (1k) blocks we need to add to get to --size, 0 if we're already there or beyond. """ msger.debug("Requested partition size for %s: %d" % \ (self.mountpoint, self.size)) if not self.size: return 0 requested_blocks = self.size * 1024 msger.debug("Requested blocks %d, current_blocks %d" % \ (requested_blocks, current_blocks)) if requested_blocks > current_blocks: return requested_blocks - current_blocks else: return 0 def prepare(self, cr, cr_workdir, oe_builddir, rootfs_dir, bootimg_dir, kernel_dir, native_sysroot): """ Prepare content for individual partitions, depending on partition command parameters. """ if not self.source: if self.fstype and self.fstype == "swap": self.prepare_swap_partition(cr_workdir, oe_builddir, native_sysroot) elif self.fstype: self.prepare_empty_partition(cr_workdir, oe_builddir, native_sysroot) return self._source_methods = pluginmgr.get_source_plugin_methods(self.source, partition_methods) self._source_methods["do_configure_partition"](self, cr, cr_workdir, oe_builddir, bootimg_dir, kernel_dir, native_sysroot) self._source_methods["do_stage_partition"](self, cr, cr_workdir, oe_builddir, bootimg_dir, kernel_dir, native_sysroot) self._source_methods["do_prepare_partition"](self, cr, cr_workdir, oe_builddir, bootimg_dir, kernel_dir, rootfs_dir, native_sysroot) def prepare_rootfs_from_fs_image(self, cr_workdir, oe_builddir, rootfs_dir): """ Handle an already-created partition e.g. xxx.ext3 """ rootfs = oe_builddir du_cmd = "du -Lbms %s" % rootfs rc, out = exec_cmd(du_cmd) rootfs_size = out.split()[0] self.size = rootfs_size self.source_file = rootfs def prepare_rootfs(self, cr_workdir, oe_builddir, rootfs_dir, native_sysroot): """ Prepare content for a rootfs partition i.e. create a partition and fill it from a /rootfs dir. Currently handles ext2/3/4 and btrfs. """ pseudo = "export PSEUDO_PREFIX=%s/usr;" % native_sysroot pseudo += "export PSEUDO_LOCALSTATEDIR=%s/../pseudo;" % rootfs_dir pseudo += "export PSEUDO_PASSWD=%s;" % rootfs_dir pseudo += "export PSEUDO_NOSYMLINKEXP=1;" pseudo += "%s/usr/bin/pseudo " % native_sysroot if self.fstype.startswith("ext"): return self.prepare_rootfs_ext(cr_workdir, oe_builddir, rootfs_dir, native_sysroot, pseudo) elif self.fstype.startswith("btrfs"): return self.prepare_rootfs_btrfs(cr_workdir, oe_builddir, rootfs_dir, native_sysroot, pseudo) def prepare_rootfs_ext(self, cr_workdir, oe_builddir, rootfs_dir, native_sysroot, pseudo): """ Prepare content for an ext2/3/4 rootfs partition. """ image_rootfs = rootfs_dir rootfs = "%s/rootfs_%s.%s" % (cr_workdir, self.label ,self.fstype) du_cmd = "du -ks %s" % image_rootfs rc, out = exec_cmd(du_cmd) actual_rootfs_size = int(out.split()[0]) extra_blocks = self.get_extra_block_count(actual_rootfs_size) if extra_blocks < IMAGE_EXTRA_SPACE: extra_blocks = IMAGE_EXTRA_SPACE rootfs_size = actual_rootfs_size + extra_blocks msger.debug("Added %d extra blocks to %s to get to %d total blocks" % \ (extra_blocks, self.mountpoint, rootfs_size)) dd_cmd = "dd if=/dev/zero of=%s bs=1024 seek=%d count=0 bs=1k" % \ (rootfs, rootfs_size) rc, out = exec_cmd(dd_cmd) extra_imagecmd = "-i 8192" mkfs_cmd = "mkfs.%s -F %s %s -d %s" % \ (self.fstype, extra_imagecmd, rootfs, image_rootfs) rc, out = exec_native_cmd(pseudo + mkfs_cmd, native_sysroot) # get the rootfs size in the right units for kickstart (Mb) du_cmd = "du -Lbms %s" % rootfs rc, out = exec_cmd(du_cmd) rootfs_size = out.split()[0] self.size = rootfs_size self.source_file = rootfs return 0 def prepare_rootfs_btrfs(self, cr_workdir, oe_builddir, rootfs_dir, native_sysroot, pseudo): """ Prepare content for a btrfs rootfs partition. Currently handles ext2/3/4 and btrfs. """ image_rootfs = rootfs_dir rootfs = "%s/rootfs_%s.%s" % (cr_workdir, self.label, self.fstype) du_cmd = "du -ks %s" % image_rootfs rc, out = exec_cmd(du_cmd) actual_rootfs_size = int(out.split()[0]) extra_blocks = self.get_extra_block_count(actual_rootfs_size) if extra_blocks < IMAGE_EXTRA_SPACE: extra_blocks = IMAGE_EXTRA_SPACE rootfs_size = actual_rootfs_size + extra_blocks msger.debug("Added %d extra blocks to %s to get to %d total blocks" % \ (extra_blocks, self.mountpoint, rootfs_size)) dd_cmd = "dd if=/dev/zero of=%s bs=1024 seek=%d count=0 bs=1k" % \ (rootfs, rootfs_size) rc, out = exec_cmd(dd_cmd) mkfs_cmd = "mkfs.%s -b %d -r %s %s" % \ (self.fstype, rootfs_size * 1024, image_rootfs, rootfs) rc, out = exec_native_cmd(pseudo + mkfs_cmd, native_sysroot) # get the rootfs size in the right units for kickstart (Mb) du_cmd = "du -Lbms %s" % rootfs rc, out = exec_cmd(du_cmd) rootfs_size = out.split()[0] self.size = rootfs_size self.source_file = rootfs def prepare_empty_partition(self, cr_workdir, oe_builddir, native_sysroot): """ Prepare an empty partition. """ if self.fstype.startswith("ext"): return self.prepare_empty_partition_ext(cr_workdir, oe_builddir, native_sysroot) elif self.fstype.startswith("btrfs"): return self.prepare_empty_partition_btrfs(cr_workdir, oe_builddir, native_sysroot) def prepare_empty_partition_ext(self, cr_workdir, oe_builddir, native_sysroot): """ Prepare an empty ext2/3/4 partition. """ fs = "%s/fs.%s" % (cr_workdir, self.fstype) dd_cmd = "dd if=/dev/zero of=%s bs=1M seek=%d count=0" % \ (fs, self.size) rc, out = exec_cmd(dd_cmd) extra_imagecmd = "-i 8192" mkfs_cmd = "mkfs.%s -F %s %s" % (self.fstype, extra_imagecmd, fs) rc, out = exec_native_cmd(mkfs_cmd, native_sysroot) self.source_file = fs return 0 def prepare_empty_partition_btrfs(self, cr_workdir, oe_builddir, native_sysroot): """ Prepare an empty btrfs partition. """ fs = "%s/fs.%s" % (cr_workdir, self.fstype) dd_cmd = "dd if=/dev/zero of=%s bs=1M seek=%d count=0" % \ (fs, self.size) rc, out = exec_cmd(dd_cmd) mkfs_cmd = "mkfs.%s -b %d %s" % (self.fstype, self.size * 1024, rootfs) rc, out = exec_native_cmd(mkfs_cmd, native_sysroot) mkfs_cmd = "mkfs.%s -F %s %s" % (self.fstype, extra_imagecmd, fs) rc, out = exec_native_cmd(mkfs_cmd, native_sysroot) self.source_file = fs return 0 def prepare_swap_partition(self, cr_workdir, oe_builddir, native_sysroot): """ Prepare a swap partition. """ fs = "%s/fs.%s" % (cr_workdir, self.fstype) dd_cmd = "dd if=/dev/zero of=%s bs=1M seek=%d count=0" % \ (fs, self.size) rc, out = exec_cmd(dd_cmd) import uuid label_str = "" if self.label: label_str = "-L %s" % self.label mkswap_cmd = "mkswap %s -U %s %s" % (label_str, str(uuid.uuid1()), fs) rc, out = exec_native_cmd(mkswap_cmd, native_sysroot) self.source_file = fs return 0 class Wic_Partition(Mic_Partition): removedKeywords = Mic_Partition.removedKeywords removedAttrs = Mic_Partition.removedAttrs def _getParser(self): op = Mic_Partition._getParser(self) # use specified source file to fill the partition # and calculate partition size op.add_option("--source", type="string", action="store", dest="source", default=None) # use specified rootfs path to fill the partition op.add_option("--rootfs-dir", type="string", action="store", dest="rootfs", default=None) return op ########NEW FILE######## __FILENAME__ = wicboot # ex:ts=4:sw=4:sts=4:et # -*- tab-width: 4; c-basic-offset: 4; indent-tabs-mode: nil -*- # # Copyright (c) 2014, Intel Corporation. # All rights reserved. # # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License version 2 as # published by the Free Software Foundation. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., # 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA. # # DESCRIPTION # This module provides the OpenEmbedded bootloader object definitions. # # AUTHORS # Tom Zanussi <tom.zanussi (at] linux.intel.com> # from pykickstart.base import * from pykickstart.errors import * from pykickstart.options import * from pykickstart.commands.bootloader import * from mic.kickstart.custom_commands.micboot import * class Wic_Bootloader(Mic_Bootloader): def __init__(self, writePriority=10, appendLine="", driveorder=None, forceLBA=False, location="", md5pass="", password="", upgrade=False, menus=""): Mic_Bootloader.__init__(self, writePriority, appendLine, driveorder, forceLBA, location, md5pass, password, upgrade) self.source = "" def _getArgsAsStr(self): retval = Mic_Bootloader._getArgsAsStr(self) if self.source: retval += " --source=%s" % self.source return retval def _getParser(self): op = Mic_Bootloader._getParser(self) # use specified source plugin to implement bootloader-specific methods op.add_option("--source", type="string", action="store", dest="source", default=None) return op ########NEW FILE######## __FILENAME__ = msger #!/usr/bin/python -tt # vim: ai ts=4 sts=4 et sw=4 # # Copyright (c) 2009, 2010, 2011 Intel, Inc. # # This program is free software; you can redistribute it and/or modify it # under the terms of the GNU General Public License as published by the Free # Software Foundation; version 2 of the License # # This program is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY # or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License # for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., 59 # Temple Place - Suite 330, Boston, MA 02111-1307, USA. import os,sys import re import time __ALL__ = ['set_mode', 'get_loglevel', 'set_loglevel', 'set_logfile', 'raw', 'debug', 'verbose', 'info', 'warning', 'error', 'ask', 'pause', ] # COLORs in ANSI INFO_COLOR = 32 # green WARN_COLOR = 33 # yellow ERR_COLOR = 31 # red ASK_COLOR = 34 # blue NO_COLOR = 0 PREFIX_RE = re.compile('^<(.*?)>\s*(.*)', re.S) INTERACTIVE = True LOG_LEVEL = 1 LOG_LEVELS = { 'quiet': 0, 'normal': 1, 'verbose': 2, 'debug': 3, 'never': 4, } LOG_FILE_FP = None LOG_CONTENT = '' CATCHERR_BUFFILE_FD = -1 CATCHERR_BUFFILE_PATH = None CATCHERR_SAVED_2 = -1 def _general_print(head, color, msg = None, stream = None, level = 'normal'): global LOG_CONTENT if not stream: stream = sys.stdout if LOG_LEVELS[level] > LOG_LEVEL: # skip return # encode raw 'unicode' str to utf8 encoded str if msg and isinstance(msg, unicode): msg = msg.encode('utf-8', 'ignore') errormsg = '' if CATCHERR_BUFFILE_FD > 0: size = os.lseek(CATCHERR_BUFFILE_FD , 0, os.SEEK_END) os.lseek(CATCHERR_BUFFILE_FD, 0, os.SEEK_SET) errormsg = os.read(CATCHERR_BUFFILE_FD, size) os.ftruncate(CATCHERR_BUFFILE_FD, 0) # append error msg to LOG if errormsg: LOG_CONTENT += errormsg # append normal msg to LOG save_msg = msg.strip() if msg else None if save_msg: timestr = time.strftime("[%m/%d %H:%M:%S %Z] ", time.localtime()) LOG_CONTENT += timestr + save_msg + '\n' if errormsg: _color_print('', NO_COLOR, errormsg, stream, level) _color_print(head, color, msg, stream, level) def _color_print(head, color, msg, stream, level): colored = True if color == NO_COLOR or \ not stream.isatty() or \ os.getenv('ANSI_COLORS_DISABLED') is not None: colored = False if head.startswith('\r'): # need not \n at last newline = False else: newline = True if colored: head = '\033[%dm%s:\033[0m ' %(color, head) if not newline: # ESC cmd to clear line head = '\033[2K' + head else: if head: head += ': ' if head.startswith('\r'): head = head.lstrip() newline = True if msg is not None: if isinstance(msg, unicode): msg = msg.encode('utf8', 'ignore') stream.write('%s%s' % (head, msg)) if newline: stream.write('\n') stream.flush() def _color_perror(head, color, msg, level = 'normal'): if CATCHERR_BUFFILE_FD > 0: _general_print(head, color, msg, sys.stdout, level) else: _general_print(head, color, msg, sys.stderr, level) def _split_msg(head, msg): if isinstance(msg, list): msg = '\n'.join(map(str, msg)) if msg.startswith('\n'): # means print \n at first msg = msg.lstrip() head = '\n' + head elif msg.startswith('\r'): # means print \r at first msg = msg.lstrip() head = '\r' + head m = PREFIX_RE.match(msg) if m: head += ' <%s>' % m.group(1) msg = m.group(2) return head, msg def get_loglevel(): return (k for k,v in LOG_LEVELS.items() if v==LOG_LEVEL).next() def set_loglevel(level): global LOG_LEVEL if level not in LOG_LEVELS: # no effect return LOG_LEVEL = LOG_LEVELS[level] def set_interactive(mode=True): global INTERACTIVE if mode: INTERACTIVE = True else: INTERACTIVE = False def log(msg=''): # log msg to LOG_CONTENT then save to logfile global LOG_CONTENT if msg: LOG_CONTENT += msg def raw(msg=''): _general_print('', NO_COLOR, msg) def info(msg): head, msg = _split_msg('Info', msg) _general_print(head, INFO_COLOR, msg) def verbose(msg): head, msg = _split_msg('Verbose', msg) _general_print(head, INFO_COLOR, msg, level = 'verbose') def warning(msg): head, msg = _split_msg('Warning', msg) _color_perror(head, WARN_COLOR, msg) def debug(msg): head, msg = _split_msg('Debug', msg) _color_perror(head, ERR_COLOR, msg, level = 'debug') def error(msg): head, msg = _split_msg('Error', msg) _color_perror(head, ERR_COLOR, msg) sys.exit(1) def ask(msg, default=True): _general_print('\rQ', ASK_COLOR, '') try: if default: msg += '(Y/n) ' else: msg += '(y/N) ' if INTERACTIVE: while True: repl = raw_input(msg) if repl.lower() == 'y': return True elif repl.lower() == 'n': return False elif not repl.strip(): # <Enter> return default # else loop else: if default: msg += ' Y' else: msg += ' N' _general_print('', NO_COLOR, msg) return default except KeyboardInterrupt: sys.stdout.write('\n') sys.exit(2) def choice(msg, choices, default=0): if default >= len(choices): return None _general_print('\rQ', ASK_COLOR, '') try: msg += " [%s] " % '/'.join(choices) if INTERACTIVE: while True: repl = raw_input(msg) if repl in choices: return repl elif not repl.strip(): return choices[default] else: msg += choices[default] _general_print('', NO_COLOR, msg) return choices[default] except KeyboardInterrupt: sys.stdout.write('\n') sys.exit(2) def pause(msg=None): if INTERACTIVE: _general_print('\rQ', ASK_COLOR, '') if msg is None: msg = 'press <ENTER> to continue ...' raw_input(msg) def set_logfile(fpath): global LOG_FILE_FP def _savelogf(): if LOG_FILE_FP: fp = open(LOG_FILE_FP, 'w') fp.write(LOG_CONTENT) fp.close() if LOG_FILE_FP is not None: warning('duplicate log file configuration') LOG_FILE_FP = fpath import atexit atexit.register(_savelogf) def enable_logstderr(fpath): global CATCHERR_BUFFILE_FD global CATCHERR_BUFFILE_PATH global CATCHERR_SAVED_2 if os.path.exists(fpath): os.remove(fpath) CATCHERR_BUFFILE_PATH = fpath CATCHERR_BUFFILE_FD = os.open(CATCHERR_BUFFILE_PATH, os.O_RDWR|os.O_CREAT) CATCHERR_SAVED_2 = os.dup(2) os.dup2(CATCHERR_BUFFILE_FD, 2) def disable_logstderr(): global CATCHERR_BUFFILE_FD global CATCHERR_BUFFILE_PATH global CATCHERR_SAVED_2 raw(msg = None) # flush message buffer and print it. os.dup2(CATCHERR_SAVED_2, 2) os.close(CATCHERR_SAVED_2) os.close(CATCHERR_BUFFILE_FD) os.unlink(CATCHERR_BUFFILE_PATH) CATCHERR_BUFFILE_FD = -1 CATCHERR_BUFFILE_PATH = None CATCHERR_SAVED_2 = -1 ########NEW FILE######## __FILENAME__ = plugin #!/usr/bin/python -tt # # Copyright (c) 2011 Intel, Inc. # # This program is free software; you can redistribute it and/or modify it # under the terms of the GNU General Public License as published by the Free # Software Foundation; version 2 of the License # # This program is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY # or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License # for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., 59 # Temple Place - Suite 330, Boston, MA 02111-1307, USA. import os, sys from mic import msger from mic import pluginbase from mic.utils import errors from mic.utils.oe.misc import * __ALL__ = ['PluginMgr', 'pluginmgr'] PLUGIN_TYPES = ["imager", "source"] # TODO "hook" PLUGIN_DIR = "/lib/mic/plugins" # relative to scripts SCRIPTS_PLUGIN_DIR = "scripts" + PLUGIN_DIR class PluginMgr(object): plugin_dirs = {} # make the manager class as singleton _instance = None def __new__(cls, *args, **kwargs): if not cls._instance: cls._instance = super(PluginMgr, cls).__new__(cls, *args, **kwargs) return cls._instance def __init__(self): mic_path = os.path.dirname(__file__) eos = mic_path.find('scripts') + len('scripts') scripts_path = mic_path[:eos] self.scripts_path = scripts_path self.plugin_dir = scripts_path + PLUGIN_DIR self.layers_path = None def _build_plugin_dir_list(self, dl, ptype): if self.layers_path is None: self.layers_path = get_bitbake_var("BBLAYERS") layer_dirs = [] for layer_path in self.layers_path.split(): path = os.path.join(layer_path, SCRIPTS_PLUGIN_DIR, ptype) layer_dirs.append(path) path = os.path.join(dl, ptype) layer_dirs.append(path) return layer_dirs def append_dirs(self, dirs): for path in dirs: self._add_plugindir(path) # load all the plugins AGAIN self._load_all() def _add_plugindir(self, path): path = os.path.abspath(os.path.expanduser(path)) if not os.path.isdir(path): msger.debug("Plugin dir is not a directory or does not exist: %s"\ % path) return if path not in self.plugin_dirs: self.plugin_dirs[path] = False # the value True/False means "loaded" def _load_all(self): for (pdir, loaded) in self.plugin_dirs.iteritems(): if loaded: continue sys.path.insert(0, pdir) for mod in [x[:-3] for x in os.listdir(pdir) if x.endswith(".py")]: if mod and mod != '__init__': if mod in sys.modules: #self.plugin_dirs[pdir] = True msger.warning("Module %s already exists, skip" % mod) else: try: pymod = __import__(mod) self.plugin_dirs[pdir] = True msger.debug("Plugin module %s:%s imported"\ % (mod, pymod.__file__)) except ImportError, err: msg = 'Failed to load plugin %s/%s: %s' \ % (os.path.basename(pdir), mod, err) msger.warning(msg) del(sys.path[0]) def get_plugins(self, ptype): """ the return value is dict of name:class pairs """ if ptype not in PLUGIN_TYPES: raise errors.CreatorError('%s is not valid plugin type' % ptype) plugins_dir = self._build_plugin_dir_list(self.plugin_dir, ptype) self.append_dirs(plugins_dir) return pluginbase.get_plugins(ptype) def get_source_plugin_methods(self, source_name, methods): """ The methods param is a dict with the method names to find. On return, the dict values will be filled in with pointers to the corresponding methods. If one or more methods are not found, None is returned. """ return_methods = None for _source_name, klass in self.get_plugins('source').iteritems(): if _source_name == source_name: for _method_name in methods.keys(): if not hasattr(klass, _method_name): msger.warning("Unimplemented %s source interface for: %s"\ % (_method_name, _source_name)) return None func = getattr(klass, _method_name) methods[_method_name] = func return_methods = methods return return_methods pluginmgr = PluginMgr() ########NEW FILE######## __FILENAME__ = pluginbase #!/usr/bin/python -tt # # Copyright (c) 2011 Intel, Inc. # # This program is free software; you can redistribute it and/or modify it # under the terms of the GNU General Public License as published by the Free # Software Foundation; version 2 of the License # # This program is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY # or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License # for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., 59 # Temple Place - Suite 330, Boston, MA 02111-1307, USA. import os import shutil from mic import msger from mic.utils import errors class _Plugin(object): class __metaclass__(type): def __init__(cls, name, bases, attrs): if not hasattr(cls, 'plugins'): cls.plugins = {} elif 'mic_plugin_type' in attrs: if attrs['mic_plugin_type'] not in cls.plugins: cls.plugins[attrs['mic_plugin_type']] = {} elif hasattr(cls, 'mic_plugin_type') and 'name' in attrs: cls.plugins[cls.mic_plugin_type][attrs['name']] = cls def show_plugins(cls): for cls in cls.plugins[cls.mic_plugin_type]: print cls def get_plugins(cls): return cls.plugins class ImagerPlugin(_Plugin): mic_plugin_type = "imager" @classmethod def check_image_exists(self, destdir, apacking=None, images=(), release=None): # if it's a packing file, reset images if apacking: images = [apacking] # release option will override images if release is not None: images = [os.path.basename(destdir.rstrip('/'))] destdir = os.path.dirname(destdir.rstrip('/')) for name in images: if not name: continue image = os.path.join(destdir, name) if not os.path.exists(image): continue if msger.ask("Target image/dir: %s already exists, " "clean up and continue?" % image): if os.path.isdir(image): shutil.rmtree(image) else: os.unlink(image) else: raise errors.Abort("Cancled") def do_create(self): pass def do_chroot(self): pass class SourcePlugin(_Plugin): mic_plugin_type = "source" """ The methods that can be implemented by --source plugins. Any methods not implemented in a subclass inherit these. """ @classmethod def do_install_disk(self, disk, disk_name, cr, workdir, oe_builddir, bootimg_dir, kernel_dir, native_sysroot): """ Called after all partitions have been prepared and assembled into a disk image. This provides a hook to allow finalization of a disk image e.g. to write an MBR to it. """ msger.debug("SourcePlugin: do_install_disk: disk: %s" % disk_name) @classmethod def do_stage_partition(self, part, cr, workdir, oe_builddir, bootimg_dir, kernel_dir, native_sysroot): """ Special content staging hook called before do_prepare_partition(), normally empty. Typically, a partition will just use the passed-in parame e.g straight bootimg_dir, etc, but in some cases, things need to be more tailored e.g. to use a deploy dir + /boot, etc. This hook allows those files to be staged in a customized fashion. Not that get_bitbake_var() allows you to acces non-standard variables that you might want to use for this. """ msger.debug("SourcePlugin: do_stage_partition: part: %s" % part) @classmethod def do_configure_partition(self, part, cr, cr_workdir, oe_builddir, bootimg_dir, kernel_dir, native_sysroot): """ Called before do_prepare_partition(), typically used to create custom configuration files for a partition, for example syslinux or grub config files. """ msger.debug("SourcePlugin: do_configure_partition: part: %s" % part) @classmethod def do_prepare_partition(self, part, cr, cr_workdir, oe_builddir, bootimg_dir, kernel_dir, rootfs_dir, native_sysroot): """ Called to do the actual content population for a partition i.e. it 'prepares' the partition to be incorporated into the image. """ msger.debug("SourcePlugin: do_prepare_partition: part: %s" % part) class BackendPlugin(_Plugin): mic_plugin_type="backend" def addRepository(self): pass def get_plugins(typen): ps = ImagerPlugin.get_plugins() if typen in ps: return ps[typen] else: return None __all__ = ['ImagerPlugin', 'BackendPlugin', 'SourcePlugin', 'get_plugins'] ########NEW FILE######## __FILENAME__ = yumpkgmgr #!/usr/bin/python -tt # # Copyright (c) 2007 Red Hat Inc. # Copyright (c) 2010, 2011 Intel, Inc. # # This program is free software; you can redistribute it and/or modify it # under the terms of the GNU General Public License as published by the Free # Software Foundation; version 2 of the License # # This program is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY # or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License # for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., 59 # Temple Place - Suite 330, Boston, MA 02111-1307, USA. import os, sys import re import tempfile import glob from string import Template import rpmUtils import yum from mic import msger from mic.kickstart import ksparser from mic.utils import misc, rpmmisc from mic.utils.grabber import TextProgress from mic.utils.proxy import get_proxy_for from mic.utils.errors import CreatorError from mic.imager.baseimager import BaseImageCreator YUMCONF_TEMP = """[main] installroot=$installroot cachedir=/var/cache/yum persistdir=/var/lib/yum plugins=0 reposdir= failovermethod=priority http_caching=packages sslverify=1 """ class MyYumRepository(yum.yumRepo.YumRepository): def __del__(self): pass def dirSetup(self): super(MyYumRepository, self).dirSetup() # relocate package dir pkgdir = os.path.join(self.basecachedir, 'packages', self.id) self.setAttribute('_dir_setup_pkgdir', pkgdir) self._dirSetupMkdir_p(self.pkgdir) def _getFile(self, url=None, relative=None, local=None, start=None, end=None, copy_local=None, checkfunc=None, text=None, reget='simple', cache=True, size=None): m2c_connection = None if not self.sslverify: try: import M2Crypto m2c_connection = M2Crypto.SSL.Connection.clientPostConnectionCheck M2Crypto.SSL.Connection.clientPostConnectionCheck = None except ImportError, err: raise CreatorError("%s, please try to install python-m2crypto" % str(err)) proxy = None if url: proxy = get_proxy_for(url) else: proxy = get_proxy_for(self.urls[0]) if proxy: self.proxy = str(proxy) size = int(size) if size else None rvalue = super(MyYumRepository, self)._getFile(url, relative, local, start, end, copy_local, checkfunc, text, reget, cache, size) if m2c_connection and \ not M2Crypto.SSL.Connection.clientPostConnectionCheck: M2Crypto.SSL.Connection.clientPostConnectionCheck = m2c_connection return rvalue from mic.pluginbase import BackendPlugin class Yum(BackendPlugin, yum.YumBase): name = 'yum' def __init__(self, target_arch, instroot, cachedir): yum.YumBase.__init__(self) self.cachedir = cachedir self.instroot = instroot self.target_arch = target_arch if self.target_arch: if not rpmUtils.arch.arches.has_key(self.target_arch): rpmUtils.arch.arches["armv7hl"] = "noarch" rpmUtils.arch.arches["armv7tnhl"] = "armv7nhl" rpmUtils.arch.arches["armv7tnhl"] = "armv7thl" rpmUtils.arch.arches["armv7thl"] = "armv7hl" rpmUtils.arch.arches["armv7nhl"] = "armv7hl" self.arch.setup_arch(self.target_arch) self.__pkgs_license = {} self.__pkgs_content = {} self.__pkgs_vcsinfo = {} self.install_debuginfo = False def doFileLogSetup(self, uid, logfile): # don't do the file log for the livecd as it can lead to open fds # being left and an inability to clean up after ourself pass def close(self): try: os.unlink(self.confpath) os.unlink(self.conf.installroot + "/yum.conf") except: pass if self.ts: self.ts.close() self._delRepos() self._delSacks() yum.YumBase.close(self) self.closeRpmDB() if not os.path.exists("/etc/fedora-release") and \ not os.path.exists("/etc/meego-release"): for i in range(3, os.sysconf("SC_OPEN_MAX")): try: os.close(i) except: pass def __del__(self): pass def _writeConf(self, confpath, installroot): conf = Template(YUMCONF_TEMP).safe_substitute(installroot=installroot) f = file(confpath, "w+") f.write(conf) f.close() os.chmod(confpath, 0644) def _cleanupRpmdbLocks(self, installroot): # cleans up temporary files left by bdb so that differing # versions of rpm don't cause problems for f in glob.glob(installroot + "/var/lib/rpm/__db*"): os.unlink(f) def setup(self): # create yum.conf (fn, self.confpath) = tempfile.mkstemp(dir=self.cachedir, prefix='yum.conf-') os.close(fn) self._writeConf(self.confpath, self.instroot) self._cleanupRpmdbLocks(self.instroot) # do setup self.doConfigSetup(fn = self.confpath, root = self.instroot) self.conf.cache = 0 self.doTsSetup() self.doRpmDBSetup() self.doRepoSetup() self.doSackSetup() def preInstall(self, pkg): # FIXME: handle pre-install package return None def selectPackage(self, pkg): """Select a given package. Can be specified with name.arch or name* """ try: self.install(pattern = pkg) return None except yum.Errors.InstallError: return "No package(s) available to install" except yum.Errors.RepoError, e: raise CreatorError("Unable to download from repo : %s" % (e,)) except yum.Errors.YumBaseError, e: raise CreatorError("Unable to install: %s" % (e,)) def deselectPackage(self, pkg): """Deselect package. Can be specified as name.arch or name* """ sp = pkg.rsplit(".", 2) txmbrs = [] if len(sp) == 2: txmbrs = self.tsInfo.matchNaevr(name=sp[0], arch=sp[1]) if len(txmbrs) == 0: exact, match, unmatch = yum.packages.parsePackages( self.pkgSack.returnPackages(), [pkg], casematch=1) for p in exact + match: txmbrs.append(p) if len(txmbrs) > 0: for x in txmbrs: self.tsInfo.remove(x.pkgtup) # we also need to remove from the conditionals # dict so that things don't get pulled back in as a result # of them. yes, this is ugly. conditionals should die. for req, pkgs in self.tsInfo.conditionals.iteritems(): if x in pkgs: pkgs.remove(x) self.tsInfo.conditionals[req] = pkgs else: msger.warning("No such package %s to remove" %(pkg,)) def selectGroup(self, grp, include = ksparser.GROUP_DEFAULT): try: yum.YumBase.selectGroup(self, grp) if include == ksparser.GROUP_REQUIRED: for p in grp.default_packages.keys(): self.deselectPackage(p) elif include == ksparser.GROUP_ALL: for p in grp.optional_packages.keys(): self.selectPackage(p) return None except (yum.Errors.InstallError, yum.Errors.GroupsError), e: return e except yum.Errors.RepoError, e: raise CreatorError("Unable to download from repo : %s" % (e,)) except yum.Errors.YumBaseError, e: raise CreatorError("Unable to install: %s" % (e,)) def addRepository(self, name, url = None, mirrorlist = None, proxy = None, proxy_username = None, proxy_password = None, inc = None, exc = None, ssl_verify=True, nocache=False, cost = None, priority=None): # TODO: Handle priority attribute for repos def _varSubstitute(option): # takes a variable and substitutes like yum configs do option = option.replace("$basearch", rpmUtils.arch.getBaseArch()) option = option.replace("$arch", rpmUtils.arch.getCanonArch()) return option repo = MyYumRepository(name) # Set proxy repo.proxy = proxy repo.proxy_username = proxy_username repo.proxy_password = proxy_password if url: repo.baseurl.append(_varSubstitute(url)) # check LICENSE files if not rpmmisc.checkRepositoryEULA(name, repo): msger.warning('skip repo:%s for failed EULA confirmation' % name) return None if mirrorlist: repo.mirrorlist = _varSubstitute(mirrorlist) conf = yum.config.RepoConf() for k, v in conf.iteritems(): if v or not hasattr(repo, k): repo.setAttribute(k, v) repo.sslverify = ssl_verify repo.cache = not nocache repo.basecachedir = self.cachedir repo.base_persistdir = self.conf.persistdir repo.failovermethod = "priority" repo.metadata_expire = 0 # Enable gpg check for verifying corrupt packages repo.gpgcheck = 1 repo.enable() repo.setup(0) self.repos.add(repo) if cost: repo.cost = cost msger.verbose('repo: %s was added' % name) return repo def installLocal(self, pkg, po=None, updateonly=False): ts = rpmUtils.transaction.initReadOnlyTransaction() try: hdr = rpmUtils.miscutils.hdrFromPackage(ts, pkg) except rpmUtils.RpmUtilsError, e: raise yum.Errors.MiscError, \ 'Could not open local rpm file: %s: %s' % (pkg, e) self.deselectPackage(hdr['name']) yum.YumBase.installLocal(self, pkg, po, updateonly) def installHasFile(self, file): provides_pkg = self.whatProvides(file, None, None) dlpkgs = map( lambda x: x.po, filter( lambda txmbr: txmbr.ts_state in ("i", "u"), self.tsInfo.getMembers())) for p in dlpkgs: for q in provides_pkg: if (p == q): return True return False def runInstall(self, checksize = 0): os.environ["HOME"] = "/" os.environ["LD_PRELOAD"] = "" try: (res, resmsg) = self.buildTransaction() except yum.Errors.RepoError, e: raise CreatorError("Unable to download from repo : %s" %(e,)) if res != 2: raise CreatorError("Failed to build transaction : %s" \ % str.join("\n", resmsg)) dlpkgs = map( lambda x: x.po, filter( lambda txmbr: txmbr.ts_state in ("i", "u"), self.tsInfo.getMembers())) # record all pkg and the content for pkg in dlpkgs: pkg_long_name = misc.RPM_FMT % { 'name': pkg.name, 'arch': pkg.arch, 'version': pkg.version, 'release': pkg.release } self.__pkgs_content[pkg_long_name] = pkg.files license = pkg.license if license in self.__pkgs_license.keys(): self.__pkgs_license[license].append(pkg_long_name) else: self.__pkgs_license[license] = [pkg_long_name] total_count = len(dlpkgs) cached_count = 0 download_total_size = sum(map(lambda x: int(x.packagesize), dlpkgs)) msger.info("\nChecking packages cached ...") for po in dlpkgs: local = po.localPkg() repo = filter(lambda r: r.id == po.repoid, self.repos.listEnabled())[0] if not repo.cache and os.path.exists(local): os.unlink(local) if not os.path.exists(local): continue if not self.verifyPkg(local, po, False): msger.warning("Package %s is damaged: %s" \ % (os.path.basename(local), local)) else: download_total_size -= int(po.packagesize) cached_count +=1 cache_avail_size = misc.get_filesystem_avail(self.cachedir) if cache_avail_size < download_total_size: raise CreatorError("No enough space used for downloading.") # record the total size of installed pkgs pkgs_total_size = 0L for x in dlpkgs: if hasattr(x, 'installedsize'): pkgs_total_size += int(x.installedsize) else: pkgs_total_size += int(x.size) # check needed size before actually download and install if checksize and pkgs_total_size > checksize: raise CreatorError("No enough space used for installing, " "please resize partition size in ks file") msger.info("Packages: %d Total, %d Cached, %d Missed" \ % (total_count, cached_count, total_count - cached_count)) try: repos = self.repos.listEnabled() for repo in repos: repo.setCallback(TextProgress(total_count - cached_count)) self.downloadPkgs(dlpkgs) # FIXME: sigcheck? self.initActionTs() self.populateTs(keepold=0) deps = self.ts.check() if len(deps) != 0: # This isn't fatal, Ubuntu has this issue but it is ok. msger.debug(deps) msger.warning("Dependency check failed!") rc = self.ts.order() if rc != 0: raise CreatorError("ordering packages for installation failed") # FIXME: callback should be refactored a little in yum cb = rpmmisc.RPMInstallCallback(self.ts) cb.tsInfo = self.tsInfo cb.filelog = False msger.warning('\nCaution, do NOT interrupt the installation, ' 'else mic cannot finish the cleanup.') installlogfile = "%s/__catched_stderr.buf" % (self.instroot) msger.enable_logstderr(installlogfile) self.runTransaction(cb) self._cleanupRpmdbLocks(self.conf.installroot) except rpmUtils.RpmUtilsError, e: raise CreatorError("mic does NOT support delta rpm: %s" % e) except yum.Errors.RepoError, e: raise CreatorError("Unable to download from repo : %s" % e) except yum.Errors.YumBaseError, e: raise CreatorError("Unable to install: %s" % e) finally: msger.disable_logstderr() def getVcsInfo(self): return self.__pkgs_vcsinfo def getAllContent(self): return self.__pkgs_content def getPkgsLicense(self): return self.__pkgs_license def getFilelist(self, pkgname): if not pkgname: return None pkg = filter(lambda txmbr: txmbr.po.name == pkgname, self.tsInfo.getMembers()) if not pkg: return None return pkg[0].po.filelist def package_url(self, pkgname): pkgs = self.pkgSack.searchNevra(name=pkgname) if pkgs: proxy = None proxies = None url = pkgs[0].remote_url repoid = pkgs[0].repoid repos = filter(lambda r: r.id == repoid, self.repos.listEnabled()) if repos: proxy = repos[0].proxy if not proxy: proxy = get_proxy_for(url) if proxy: proxies = {str(url.split(':')[0]): str(proxy)} return (url, proxies) return (None, None) ########NEW FILE######## __FILENAME__ = zypppkgmgr #!/usr/bin/python -tt # # Copyright (c) 2010, 2011 Intel, Inc. # # This program is free software; you can redistribute it and/or modify it # under the terms of the GNU General Public License as published by the Free # Software Foundation; version 2 of the License # # This program is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY # or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License # for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., 59 # Temple Place - Suite 330, Boston, MA 02111-1307, USA. import os import shutil import urlparse import rpm import zypp if not hasattr(zypp, 'PoolQuery') or \ not hasattr(zypp.RepoManager, 'loadSolvFile'): raise ImportError("python-zypp in host system cannot support PoolQuery or " "loadSolvFile interface, please update it to enhanced " "version which can be found in download.tizen.org/tools") from mic import msger from mic.kickstart import ksparser from mic.utils import misc, rpmmisc, runner, fs_related from mic.utils.grabber import myurlgrab, TextProgress from mic.utils.proxy import get_proxy_for from mic.utils.errors import CreatorError, RepoError, RpmError from mic.imager.baseimager import BaseImageCreator class RepositoryStub: def __init__(self): self.name = None self.baseurl = [] self.mirrorlist = None self.proxy = None self.proxy_username = None self.proxy_password = None self.nocache = False self.enabled = True self.autorefresh = True self.keeppackages = True self.priority = None from mic.pluginbase import BackendPlugin class Zypp(BackendPlugin): name = 'zypp' def __init__(self, target_arch, instroot, cachedir): self.cachedir = cachedir self.instroot = instroot self.target_arch = target_arch self.__pkgs_license = {} self.__pkgs_content = {} self.__pkgs_vcsinfo = {} self.repos = [] self.to_deselect = [] self.localpkgs = {} self.repo_manager = None self.repo_manager_options = None self.Z = None self.ts = None self.ts_pre = None self.incpkgs = {} self.excpkgs = {} self.pre_pkgs = [] self.probFilterFlags = [ rpm.RPMPROB_FILTER_OLDPACKAGE, rpm.RPMPROB_FILTER_REPLACEPKG ] self.has_prov_query = True self.install_debuginfo = False def doFileLogSetup(self, uid, logfile): # don't do the file log for the livecd as it can lead to open fds # being left and an inability to clean up after ourself pass def closeRpmDB(self): pass def close(self): if self.ts: self.ts.closeDB() self.ts = None if self.ts_pre: self.ts_pre.closeDB() self.ts = None self.closeRpmDB() if not os.path.exists("/etc/fedora-release") and \ not os.path.exists("/etc/meego-release"): for i in range(3, os.sysconf("SC_OPEN_MAX")): try: os.close(i) except: pass def __del__(self): self.close() def _cleanupRpmdbLocks(self, installroot): # cleans up temporary files left by bdb so that differing # versions of rpm don't cause problems import glob for f in glob.glob(installroot + "/var/lib/rpm/__db*"): os.unlink(f) def _cleanupZyppJunk(self, installroot): try: shutil.rmtree(os.path.join(installroot, '.zypp')) except: pass def setup(self): self._cleanupRpmdbLocks(self.instroot) def whatObsolete(self, pkg): query = zypp.PoolQuery() query.addKind(zypp.ResKind.package) query.addAttribute(zypp.SolvAttr.obsoletes, pkg) query.setMatchExact() for pi in query.queryResults(self.Z.pool()): return pi return None def _zyppQueryPackage(self, pkg): query = zypp.PoolQuery() query.addKind(zypp.ResKind.package) query.addAttribute(zypp.SolvAttr.name,pkg) query.setMatchExact() for pi in query.queryResults(self.Z.pool()): return pi return None def _splitPkgString(self, pkg): sp = pkg.rsplit(".",1) name = sp[0] arch = None if len(sp) == 2: arch = sp[1] sysarch = zypp.Arch(self.target_arch) if not zypp.Arch(arch).compatible_with (sysarch): arch = None name = ".".join(sp) return name, arch def selectPackage(self, pkg): """Select a given package or package pattern, can be specified with name.arch or name* or *name """ if not self.Z: self.__initialize_zypp() def markPoolItem(obs, pi): if obs == None: pi.status().setToBeInstalled (zypp.ResStatus.USER) else: obs.status().setToBeInstalled (zypp.ResStatus.USER) def cmpEVR(p1, p2): # compare criterion: arch compatibility first, then repo # priority, and version last a1 = p1.arch() a2 = p2.arch() if str(a1) != str(a2): if a1.compatible_with(a2): return -1 else: return 1 # Priority of a repository is an integer value between 0 (the # highest priority) and 99 (the lowest priority) pr1 = int(p1.repoInfo().priority()) pr2 = int(p2.repoInfo().priority()) if pr1 > pr2: return -1 elif pr1 < pr2: return 1 ed1 = p1.edition() ed2 = p2.edition() (e1, v1, r1) = map(str, [ed1.epoch(), ed1.version(), ed1.release()]) (e2, v2, r2) = map(str, [ed2.epoch(), ed2.version(), ed2.release()]) return rpm.labelCompare((e1, v1, r1), (e2, v2, r2)) found = False startx = pkg.startswith("*") endx = pkg.endswith("*") ispattern = startx or endx name, arch = self._splitPkgString(pkg) q = zypp.PoolQuery() q.addKind(zypp.ResKind.package) if ispattern: if startx and not endx: pattern = '%s$' % (pkg[1:]) if endx and not startx: pattern = '^%s' % (pkg[0:-1]) if endx and startx: pattern = '%s' % (pkg[1:-1]) q.setMatchRegex() q.addAttribute(zypp.SolvAttr.name,pattern) elif arch: q.setMatchExact() q.addAttribute(zypp.SolvAttr.name,name) else: q.setMatchExact() q.addAttribute(zypp.SolvAttr.name,pkg) for pitem in sorted( q.queryResults(self.Z.pool()), cmp=lambda x,y: cmpEVR(zypp.asKindPackage(x), zypp.asKindPackage(y)), reverse=True): item = zypp.asKindPackage(pitem) if item.name() in self.excpkgs.keys() and \ self.excpkgs[item.name()] == item.repoInfo().name(): continue if item.name() in self.incpkgs.keys() and \ self.incpkgs[item.name()] != item.repoInfo().name(): continue found = True obspkg = self.whatObsolete(item.name()) if arch: if arch == str(item.arch()): item.status().setToBeInstalled (zypp.ResStatus.USER) else: markPoolItem(obspkg, pitem) if not ispattern: break # Can't match using package name, then search from packge # provides infomation if found == False and not ispattern: q.addAttribute(zypp.SolvAttr.provides, pkg) q.addAttribute(zypp.SolvAttr.name,'') for pitem in sorted( q.queryResults(self.Z.pool()), cmp=lambda x,y: cmpEVR(zypp.asKindPackage(x), zypp.asKindPackage(y)), reverse=True): item = zypp.asKindPackage(pitem) if item.name() in self.excpkgs.keys() and \ self.excpkgs[item.name()] == item.repoInfo().name(): continue if item.name() in self.incpkgs.keys() and \ self.incpkgs[item.name()] != item.repoInfo().name(): continue found = True obspkg = self.whatObsolete(item.name()) markPoolItem(obspkg, pitem) break if found: return None else: raise CreatorError("Unable to find package: %s" % (pkg,)) def inDeselectPackages(self, pitem): """check if specified pacakges are in the list of inDeselectPackages """ item = zypp.asKindPackage(pitem) name = item.name() for pkg in self.to_deselect: startx = pkg.startswith("*") endx = pkg.endswith("*") ispattern = startx or endx pkgname, pkgarch = self._splitPkgString(pkg) if not ispattern: if pkgarch: if name == pkgname and str(item.arch()) == pkgarch: return True; else: if name == pkgname: return True; else: if startx and name.endswith(pkg[1:]): return True; if endx and name.startswith(pkg[:-1]): return True; return False; def deselectPackage(self, pkg): """collect packages should not be installed""" self.to_deselect.append(pkg) def selectGroup(self, grp, include = ksparser.GROUP_DEFAULT): if not self.Z: self.__initialize_zypp() found = False q=zypp.PoolQuery() q.addKind(zypp.ResKind.pattern) for pitem in q.queryResults(self.Z.pool()): item = zypp.asKindPattern(pitem) summary = "%s" % item.summary() name = "%s" % item.name() if name == grp or summary == grp: found = True pitem.status().setToBeInstalled (zypp.ResStatus.USER) break if found: if include == ksparser.GROUP_REQUIRED: map( lambda p: self.deselectPackage(p), grp.default_packages.keys()) return None else: raise CreatorError("Unable to find pattern: %s" % (grp,)) def addRepository(self, name, url = None, mirrorlist = None, proxy = None, proxy_username = None, proxy_password = None, inc = None, exc = None, ssl_verify = True, nocache = False, cost=None, priority=None): # TODO: Handle cost attribute for repos if not self.repo_manager: self.__initialize_repo_manager() if not proxy and url: proxy = get_proxy_for(url) repo = RepositoryStub() repo.name = name repo.id = name repo.proxy = proxy repo.proxy_username = proxy_username repo.proxy_password = proxy_password repo.ssl_verify = ssl_verify repo.nocache = nocache repo.baseurl.append(url) if inc: for pkg in inc: self.incpkgs[pkg] = name if exc: for pkg in exc: self.excpkgs[pkg] = name # check LICENSE files if not rpmmisc.checkRepositoryEULA(name, repo): msger.warning('skip repo:%s for failed EULA confirmation' % name) return None if mirrorlist: repo.mirrorlist = mirrorlist # Enable gpg check for verifying corrupt packages repo.gpgcheck = 1 if priority is not None: # priority 0 has issue in RepoInfo.setPriority repo.priority = priority + 1 try: repo_info = zypp.RepoInfo() repo_info.setAlias(repo.name) repo_info.setName(repo.name) repo_info.setEnabled(repo.enabled) repo_info.setAutorefresh(repo.autorefresh) repo_info.setKeepPackages(repo.keeppackages) baseurl = zypp.Url(repo.baseurl[0]) if not ssl_verify: baseurl.setQueryParam("ssl_verify", "no") if proxy: scheme, host, path, parm, query, frag = urlparse.urlparse(proxy) proxyinfo = host.split(":") host = proxyinfo[0] port = "80" if len(proxyinfo) > 1: port = proxyinfo[1] if proxy.startswith("socks") and len(proxy.rsplit(':', 1)) == 2: host = proxy.rsplit(':', 1)[0] port = proxy.rsplit(':', 1)[1] baseurl.setQueryParam ("proxy", host) baseurl.setQueryParam ("proxyport", port) repo.baseurl[0] = baseurl.asCompleteString() self.repos.append(repo) repo_info.addBaseUrl(baseurl) if repo.priority is not None: repo_info.setPriority(repo.priority) # this hack is used to change zypp credential file location # the default one is $HOME/.zypp, which cause conflicts when # installing some basic packages, and the location doesn't # have any interface actually, so use a tricky way anyway homedir = None if 'HOME' in os.environ: homedir = os.environ['HOME'] os.environ['HOME'] = '/' else: os.environ['HOME'] = '/' self.repo_manager.addRepository(repo_info) # save back the $HOME env if homedir: os.environ['HOME'] = homedir else: del os.environ['HOME'] self.__build_repo_cache(name) except RuntimeError, e: raise CreatorError(str(e)) msger.verbose('repo: %s was added' % name) return repo def installHasFile(self, file): return False def preInstall(self, pkg): self.pre_pkgs.append(pkg) def runInstall(self, checksize = 0): os.environ["HOME"] = "/" os.environ["LD_PRELOAD"] = "" self.buildTransaction() todo = zypp.GetResolvablesToInsDel(self.Z.pool()) installed_pkgs = todo._toInstall dlpkgs = [] for pitem in installed_pkgs: if not zypp.isKindPattern(pitem) and \ not self.inDeselectPackages(pitem): item = zypp.asKindPackage(pitem) dlpkgs.append(item) if not self.install_debuginfo or str(item.arch()) == "noarch": continue dipkg = self._zyppQueryPackage("%s-debuginfo" % item.name()) if dipkg: ditem = zypp.asKindPackage(dipkg) dlpkgs.append(ditem) else: msger.warning("No debuginfo rpm found for: %s" \ % item.name()) # record all pkg and the content localpkgs = self.localpkgs.keys() for pkg in dlpkgs: license = '' if pkg.name() in localpkgs: hdr = rpmmisc.readRpmHeader(self.ts, self.localpkgs[pkg.name()]) pkg_long_name = misc.RPM_FMT % { 'name': hdr['name'], 'arch': hdr['arch'], 'version': hdr['version'], 'release': hdr['release'] } license = hdr['license'] else: pkg_long_name = misc.RPM_FMT % { 'name': pkg.name(), 'arch': pkg.arch(), 'version': pkg.edition().version(), 'release': pkg.edition().release() } license = pkg.license() if license in self.__pkgs_license.keys(): self.__pkgs_license[license].append(pkg_long_name) else: self.__pkgs_license[license] = [pkg_long_name] total_count = len(dlpkgs) cached_count = 0 download_total_size = sum(map(lambda x: int(x.downloadSize()), dlpkgs)) localpkgs = self.localpkgs.keys() msger.info("Checking packages cached ...") for po in dlpkgs: # Check if it is cached locally if po.name() in localpkgs: cached_count += 1 else: local = self.getLocalPkgPath(po) name = str(po.repoInfo().name()) try: repo = filter(lambda r: r.name == name, self.repos)[0] except IndexError: repo = None nocache = repo.nocache if repo else False if os.path.exists(local): if nocache or self.checkPkg(local) !=0: os.unlink(local) else: download_total_size -= int(po.downloadSize()) cached_count += 1 cache_avail_size = misc.get_filesystem_avail(self.cachedir) if cache_avail_size < download_total_size: raise CreatorError("No enough space used for downloading.") # record the total size of installed pkgs install_total_size = sum(map(lambda x: int(x.installSize()), dlpkgs)) # check needed size before actually download and install # FIXME: for multiple partitions for loop type, check fails # skip the check temporarily #if checksize and install_total_size > checksize: # raise CreatorError("No enough space used for installing, " # "please resize partition size in ks file") download_count = total_count - cached_count msger.info("Packages: %d Total, %d Cached, %d Missed" \ % (total_count, cached_count, download_count)) try: if download_count > 0: msger.info("Downloading packages ...") self.downloadPkgs(dlpkgs, download_count) self.installPkgs(dlpkgs) except (RepoError, RpmError): raise except Exception, e: raise CreatorError("Package installation failed: %s" % (e,)) def getVcsInfo(self): if self.__pkgs_vcsinfo: return if not self.ts: self.__initialize_transaction() mi = self.ts.dbMatch() for hdr in mi: lname = misc.RPM_FMT % { 'name': hdr['name'], 'arch': hdr['arch'], 'version': hdr['version'], 'release': hdr['release'] } self.__pkgs_vcsinfo[lname] = hdr['VCS'] return self.__pkgs_vcsinfo def getAllContent(self): if self.__pkgs_content: return self.__pkgs_content if not self.ts: self.__initialize_transaction() mi = self.ts.dbMatch() for hdr in mi: lname = misc.RPM_FMT % { 'name': hdr['name'], 'arch': hdr['arch'], 'version': hdr['version'], 'release': hdr['release'] } self.__pkgs_content[lname] = hdr['FILENAMES'] return self.__pkgs_content def getPkgsLicense(self): return self.__pkgs_license def getFilelist(self, pkgname): if not pkgname: return None if not self.ts: self.__initialize_transaction() mi = self.ts.dbMatch('name', pkgname) for header in mi: return header['FILENAMES'] def __initialize_repo_manager(self): if self.repo_manager: return # Clean up repo metadata shutil.rmtree(self.cachedir + "/etc", ignore_errors = True) shutil.rmtree(self.cachedir + "/solv", ignore_errors = True) shutil.rmtree(self.cachedir + "/raw", ignore_errors = True) zypp.KeyRing.setDefaultAccept( zypp.KeyRing.ACCEPT_UNSIGNED_FILE | zypp.KeyRing.ACCEPT_VERIFICATION_FAILED | zypp.KeyRing.ACCEPT_UNKNOWNKEY | zypp.KeyRing.TRUST_KEY_TEMPORARILY ) self.repo_manager_options = \ zypp.RepoManagerOptions(zypp.Pathname(self.instroot)) self.repo_manager_options.knownReposPath = \ zypp.Pathname(self.cachedir + "/etc/zypp/repos.d") self.repo_manager_options.repoCachePath = \ zypp.Pathname(self.cachedir) self.repo_manager_options.repoRawCachePath = \ zypp.Pathname(self.cachedir + "/raw") self.repo_manager_options.repoSolvCachePath = \ zypp.Pathname(self.cachedir + "/solv") self.repo_manager_options.repoPackagesCachePath = \ zypp.Pathname(self.cachedir + "/packages") self.repo_manager = zypp.RepoManager(self.repo_manager_options) def __build_repo_cache(self, name): repo = self.repo_manager.getRepositoryInfo(name) if self.repo_manager.isCached(repo) or not repo.enabled(): return msger.info('Refreshing repository: %s ...' % name) self.repo_manager.buildCache(repo, zypp.RepoManager.BuildIfNeeded) def __initialize_zypp(self): if self.Z: return zconfig = zypp.ZConfig_instance() # Set system architecture if self.target_arch: zconfig.setSystemArchitecture(zypp.Arch(self.target_arch)) msger.info("zypp architecture is <%s>" % zconfig.systemArchitecture()) # repoPackagesCachePath is corrected by this self.repo_manager = zypp.RepoManager(self.repo_manager_options) repos = self.repo_manager.knownRepositories() for repo in repos: if not repo.enabled(): continue self.repo_manager.loadFromCache(repo) self.Z = zypp.ZYppFactory_instance().getZYpp() self.Z.initializeTarget(zypp.Pathname(self.instroot)) self.Z.target().load() def buildTransaction(self): if not self.Z.resolver().resolvePool(): probs = self.Z.resolver().problems() for problem in probs: msger.warning("repo problem: %s, %s" \ % (problem.description().decode("utf-8"), problem.details().decode("utf-8"))) raise RepoError("found %d resolver problem, abort!" \ % len(probs)) def getLocalPkgPath(self, po): repoinfo = po.repoInfo() cacheroot = repoinfo.packagesPath() location= po.location() rpmpath = str(location.filename()) pkgpath = "%s/%s" % (cacheroot, os.path.basename(rpmpath)) return pkgpath def installLocal(self, pkg, po=None, updateonly=False): if not self.ts: self.__initialize_transaction() solvfile = "%s/.solv" % (self.cachedir) rc, out = runner.runtool([fs_related.find_binary_path("rpms2solv"), pkg]) if rc == 0: f = open(solvfile, "w+") f.write(out) f.close() warnmsg = self.repo_manager.loadSolvFile(solvfile, os.path.basename(pkg)) if warnmsg: msger.warning(warnmsg) os.unlink(solvfile) else: msger.warning('Can not get %s solv data.' % pkg) hdr = rpmmisc.readRpmHeader(self.ts, pkg) arch = zypp.Arch(hdr['arch']) sysarch = zypp.Arch(self.target_arch) if arch.compatible_with (sysarch): pkgname = hdr['name'] self.localpkgs[pkgname] = pkg self.selectPackage(pkgname) msger.info("Marking %s to be installed" % (pkg)) else: msger.warning("Cannot add package %s to transaction. " "Not a compatible architecture: %s" \ % (pkg, hdr['arch'])) def downloadPkgs(self, package_objects, count): localpkgs = self.localpkgs.keys() progress_obj = TextProgress(count) for po in package_objects: if po.name() in localpkgs: continue filename = self.getLocalPkgPath(po) if os.path.exists(filename): if self.checkPkg(filename) == 0: continue dirn = os.path.dirname(filename) if not os.path.exists(dirn): os.makedirs(dirn) url = self.get_url(po) proxies = self.get_proxies(po) try: filename = myurlgrab(url, filename, proxies, progress_obj) except CreatorError: self.close() raise def preinstallPkgs(self): if not self.ts_pre: self.__initialize_transaction() self.ts_pre.order() cb = rpmmisc.RPMInstallCallback(self.ts_pre) cb.headmsg = "Preinstall" installlogfile = "%s/__catched_stderr.buf" % (self.instroot) # start to catch stderr output from librpm msger.enable_logstderr(installlogfile) errors = self.ts_pre.run(cb.callback, '') # stop catch msger.disable_logstderr() self.ts_pre.closeDB() self.ts_pre = None if errors is not None: if len(errors) == 0: msger.warning('scriptlet or other non-fatal errors occurred ' 'during transaction.') else: for e in errors: msger.warning(e[0]) raise RepoError('Could not run transaction.') def installPkgs(self, package_objects): if not self.ts: self.__initialize_transaction() # clean rpm lock self._cleanupRpmdbLocks(self.instroot) self._cleanupZyppJunk(self.instroot) # Set filters probfilter = 0 for flag in self.probFilterFlags: probfilter |= flag self.ts.setProbFilter(probfilter) self.ts_pre.setProbFilter(probfilter) localpkgs = self.localpkgs.keys() for po in package_objects: pkgname = po.name() if pkgname in localpkgs: rpmpath = self.localpkgs[pkgname] else: rpmpath = self.getLocalPkgPath(po) if not os.path.exists(rpmpath): # Maybe it is a local repo rpmuri = self.get_url(po) if rpmuri.startswith("file:/"): rpmpath = rpmuri[5:] if not os.path.exists(rpmpath): raise RpmError("Error: %s doesn't exist" % rpmpath) h = rpmmisc.readRpmHeader(self.ts, rpmpath) if pkgname in self.pre_pkgs: msger.verbose("pre-install package added: %s" % pkgname) self.ts_pre.addInstall(h, rpmpath, 'u') self.ts.addInstall(h, rpmpath, 'u') unresolved_dependencies = self.ts.check() if not unresolved_dependencies: if self.pre_pkgs: self.preinstallPkgs() self.ts.order() cb = rpmmisc.RPMInstallCallback(self.ts) installlogfile = "%s/__catched_stderr.buf" % (self.instroot) # start to catch stderr output from librpm msger.enable_logstderr(installlogfile) errors = self.ts.run(cb.callback, '') # stop catch msger.disable_logstderr() self.ts.closeDB() self.ts = None if errors is not None: if len(errors) == 0: msger.warning('scriptlet or other non-fatal errors occurred ' 'during transaction.') else: for e in errors: msger.warning(e[0]) raise RepoError('Could not run transaction.') else: for pkg, need, needflags, sense, key in unresolved_dependencies: package = '-'.join(pkg) if needflags == rpm.RPMSENSE_LESS: deppkg = ' < '.join(need) elif needflags == rpm.RPMSENSE_EQUAL: deppkg = ' = '.join(need) elif needflags == rpm.RPMSENSE_GREATER: deppkg = ' > '.join(need) else: deppkg = '-'.join(need) if sense == rpm.RPMDEP_SENSE_REQUIRES: msger.warning("[%s] Requires [%s], which is not provided" \ % (package, deppkg)) elif sense == rpm.RPMDEP_SENSE_CONFLICTS: msger.warning("[%s] Conflicts with [%s]" %(package,deppkg)) raise RepoError("Unresolved dependencies, transaction failed.") def __initialize_transaction(self): if not self.ts: self.ts = rpm.TransactionSet(self.instroot) # Set to not verify DSA signatures. self.ts.setVSFlags(rpm._RPMVSF_NOSIGNATURES|rpm._RPMVSF_NODIGESTS) if not self.ts_pre: self.ts_pre = rpm.TransactionSet(self.instroot) # Just unpack the files, don't run scripts self.ts_pre.setFlags(rpm.RPMTRANS_FLAG_ALLFILES | rpm.RPMTRANS_FLAG_NOSCRIPTS) # Set to not verify DSA signatures. self.ts_pre.setVSFlags(rpm._RPMVSF_NOSIGNATURES|rpm._RPMVSF_NODIGESTS) def checkPkg(self, pkg): ret = 1 if not os.path.exists(pkg): return ret ret = rpmmisc.checkRpmIntegrity('rpm', pkg) if ret != 0: msger.warning("package %s is damaged: %s" \ % (os.path.basename(pkg), pkg)) return ret def _add_prob_flags(self, *flags): for flag in flags: if flag not in self.probFilterFlags: self.probFilterFlags.append(flag) def get_proxies(self, pobj): if not pobj: return None proxy = None proxies = None repoinfo = pobj.repoInfo() reponame = "%s" % repoinfo.name() repos = filter(lambda r: r.name == reponame, self.repos) repourl = str(repoinfo.baseUrls()[0]) if repos: proxy = repos[0].proxy if not proxy: proxy = get_proxy_for(repourl) if proxy: proxies = {str(repourl.split(':')[0]): str(proxy)} return proxies def get_url(self, pobj): if not pobj: return None name = str(pobj.repoInfo().name()) try: repo = filter(lambda r: r.name == name, self.repos)[0] except IndexError: return None baseurl = repo.baseurl[0] index = baseurl.find("?") if index > -1: baseurl = baseurl[:index] location = pobj.location() location = str(location.filename()) if location.startswith("./"): location = location[2:] return os.path.join(baseurl, location) def package_url(self, pkgname): def cmpEVR(p1, p2): ed1 = p1.edition() ed2 = p2.edition() (e1, v1, r1) = map(str, [ed1.epoch(), ed1.version(), ed1.release()]) (e2, v2, r2) = map(str, [ed2.epoch(), ed2.version(), ed2.release()]) return rpm.labelCompare((e1, v1, r1), (e2, v2, r2)) if not self.Z: self.__initialize_zypp() q = zypp.PoolQuery() q.addKind(zypp.ResKind.package) q.setMatchExact() q.addAttribute(zypp.SolvAttr.name, pkgname) items = sorted(q.queryResults(self.Z.pool()), cmp=lambda x,y: cmpEVR(zypp.asKindPackage(x), zypp.asKindPackage(y)), reverse=True) if items: item = zypp.asKindPackage(items[0]) url = self.get_url(item) proxies = self.get_proxies(item) return (url, proxies) return (None, None) ########NEW FILE######## __FILENAME__ = ########NEW FILE######## __FILENAME__ = empty_hook #!/usr/bin/python # TODO: plugin base for hooks ########NEW FILE######## __FILENAME__ = direct_plugin # ex:ts=4:sw=4:sts=4:et # -*- tab-width: 4; c-basic-offset: 4; indent-tabs-mode: nil -*- # # Copyright (c) 2013, Intel Corporation. # All rights reserved. # # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License version 2 as # published by the Free Software Foundation. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., # 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA. # # DESCRIPTION # This implements the 'direct' imager plugin class for 'wic', based # loosely on the raw imager plugin from 'mic' # # AUTHORS # Tom Zanussi <tom.zanussi (at] linux.intel.com> # import os import shutil import re import tempfile from mic import chroot, msger from mic.utils import misc, fs_related, errors, runner, cmdln from mic.conf import configmgr from mic.plugin import pluginmgr from mic.utils.partitionedfs import PartitionedMount import mic.imager.direct as direct from mic.pluginbase import ImagerPlugin class DirectPlugin(ImagerPlugin): name = 'direct' @classmethod def __rootfs_dir_to_dict(self, rootfs_dirs): """ Gets a string that contain 'connection=dir' splitted by space and return a dict """ krootfs_dir = {} for rootfs_dir in rootfs_dirs.split(' '): k, v = rootfs_dir.split('=') krootfs_dir[k] = v return krootfs_dir @classmethod def do_create(self, subcmd, opts, *args): """ Create direct image, called from creator as 'direct' cmd """ if len(args) != 9: raise errors.Usage("Extra arguments given") staging_data_dir = args[0] hdddir = args[1] native_sysroot = args[2] kernel_dir = args[3] bootimg_dir = args[4] rootfs_dir = args[5] creatoropts = configmgr.create ksconf = args[6] image_output_dir = args[7] oe_builddir = args[8] krootfs_dir = self.__rootfs_dir_to_dict(rootfs_dir) configmgr._ksconf = ksconf creator = direct.DirectImageCreator(oe_builddir, image_output_dir, krootfs_dir, bootimg_dir, kernel_dir, native_sysroot, hdddir, staging_data_dir, creatoropts, None, None, None) try: creator.mount(None, creatoropts["cachedir"]) creator.install() creator.configure(creatoropts["repomd"]) creator.print_outimage_info() except errors.CreatorError: raise finally: creator.cleanup() return 0 ########NEW FILE######## __FILENAME__ = fs_plugin #!/usr/bin/python -tt # # Copyright (c) 2011 Intel, Inc. # # This program is free software; you can redistribute it and/or modify it # under the terms of the GNU General Public License as published by the Free # Software Foundation; version 2 of the License # # This program is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY # or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License # for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., 59 # Temple Place - Suite 330, Boston, MA 02111-1307, USA. import os import sys from mic import chroot, msger from mic.utils import cmdln, misc, errors, fs_related from mic.imager import fs from mic.conf import configmgr from mic.plugin import pluginmgr from mic.pluginbase import ImagerPlugin class FsPlugin(ImagerPlugin): name = 'fs' @classmethod @cmdln.option("--include-src", dest="include_src", action="store_true", default=False, help="Generate a image with source rpms included") def do_create(self, subcmd, opts, *args): """${cmd_name}: create fs image Usage: ${name} ${cmd_name} <ksfile> [OPTS] ${cmd_option_list} """ if len(args) != 1: raise errors.Usage("Extra arguments given") creatoropts = configmgr.create ksconf = args[0] if creatoropts['runtime'] == 'bootstrap': configmgr._ksconf = ksconf rt_util.bootstrap_mic() recording_pkgs = [] if len(creatoropts['record_pkgs']) > 0: recording_pkgs = creatoropts['record_pkgs'] if creatoropts['release'] is not None: if 'name' not in recording_pkgs: recording_pkgs.append('name') if 'vcs' not in recording_pkgs: recording_pkgs.append('vcs') configmgr._ksconf = ksconf # Called After setting the configmgr._ksconf as the creatoropts['name'] is reset there. if creatoropts['release'] is not None: creatoropts['outdir'] = "%s/%s/images/%s/" % (creatoropts['outdir'], creatoropts['release'], creatoropts['name']) # try to find the pkgmgr pkgmgr = None backends = pluginmgr.get_plugins('backend') if 'auto' == creatoropts['pkgmgr']: for key in configmgr.prefer_backends: if key in backends: pkgmgr = backends[key] break else: for key in backends.keys(): if key == creatoropts['pkgmgr']: pkgmgr = backends[key] break if not pkgmgr: raise errors.CreatorError("Can't find backend: %s, " "available choices: %s" % (creatoropts['pkgmgr'], ','.join(backends.keys()))) creator = fs.FsImageCreator(creatoropts, pkgmgr) creator._include_src = opts.include_src if len(recording_pkgs) > 0: creator._recording_pkgs = recording_pkgs self.check_image_exists(creator.destdir, creator.pack_to, [creator.name], creatoropts['release']) try: creator.check_depend_tools() creator.mount(None, creatoropts["cachedir"]) creator.install() #Download the source packages ###private options if opts.include_src: installed_pkgs = creator.get_installed_packages() msger.info('--------------------------------------------------') msger.info('Generating the image with source rpms included ...') if not misc.SrcpkgsDownload(installed_pkgs, creatoropts["repomd"], creator._instroot, creatoropts["cachedir"]): msger.warning("Source packages can't be downloaded") creator.configure(creatoropts["repomd"]) creator.copy_kernel() creator.unmount() creator.package(creatoropts["outdir"]) if creatoropts['release'] is not None: creator.release_output(ksconf, creatoropts['outdir'], creatoropts['release']) creator.print_outimage_info() except errors.CreatorError: raise finally: creator.cleanup() msger.info("Finished.") return 0 @classmethod def do_chroot(self, target, cmd=[]):#chroot.py parse opts&args try: if len(cmd) != 0: cmdline = ' '.join(cmd) else: cmdline = "/bin/bash" envcmd = fs_related.find_binary_inchroot("env", target) if envcmd: cmdline = "%s HOME=/root %s" % (envcmd, cmdline) chroot.chroot(target, None, cmdline) finally: chroot.cleanup_after_chroot("dir", None, None, None) return 1 ########NEW FILE######## __FILENAME__ = livecd_plugin #!/usr/bin/python -tt # # Copyright (c) 2011 Intel, Inc. # # This program is free software; you can redistribute it and/or modify it # under the terms of the GNU General Public License as published by the Free # Software Foundation; version 2 of the License # # This program is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY # or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License # for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., 59 # Temple Place - Suite 330, Boston, MA 02111-1307, USA. import os import shutil import tempfile from mic import chroot, msger from mic.utils import misc, fs_related, errors from mic.conf import configmgr import mic.imager.livecd as livecd from mic.plugin import pluginmgr from mic.pluginbase import ImagerPlugin class LiveCDPlugin(ImagerPlugin): name = 'livecd' @classmethod def do_create(self, subcmd, opts, *args): """${cmd_name}: create livecd image Usage: ${name} ${cmd_name} <ksfile> [OPTS] ${cmd_option_list} """ if len(args) != 1: raise errors.Usage("Extra arguments given") creatoropts = configmgr.create ksconf = args[0] if creatoropts['runtime'] == 'bootstrap': configmgr._ksconf = ksconf rt_util.bootstrap_mic() if creatoropts['arch'] and creatoropts['arch'].startswith('arm'): msger.warning('livecd cannot support arm images, Quit') return recording_pkgs = [] if len(creatoropts['record_pkgs']) > 0: recording_pkgs = creatoropts['record_pkgs'] if creatoropts['release'] is not None: if 'name' not in recording_pkgs: recording_pkgs.append('name') if 'vcs' not in recording_pkgs: recording_pkgs.append('vcs') configmgr._ksconf = ksconf # Called After setting the configmgr._ksconf as the creatoropts['name'] is reset there. if creatoropts['release'] is not None: creatoropts['outdir'] = "%s/%s/images/%s/" % (creatoropts['outdir'], creatoropts['release'], creatoropts['name']) # try to find the pkgmgr pkgmgr = None backends = pluginmgr.get_plugins('backend') if 'auto' == creatoropts['pkgmgr']: for key in configmgr.prefer_backends: if key in backends: pkgmgr = backends[key] break else: for key in backends.keys(): if key == creatoropts['pkgmgr']: pkgmgr = backends[key] break if not pkgmgr: raise errors.CreatorError("Can't find backend: %s, " "available choices: %s" % (creatoropts['pkgmgr'], ','.join(backends.keys()))) creator = livecd.LiveCDImageCreator(creatoropts, pkgmgr) if len(recording_pkgs) > 0: creator._recording_pkgs = recording_pkgs self.check_image_exists(creator.destdir, creator.pack_to, [creator.name + ".iso"], creatoropts['release']) try: creator.check_depend_tools() creator.mount(None, creatoropts["cachedir"]) creator.install() creator.configure(creatoropts["repomd"]) creator.copy_kernel() creator.unmount() creator.package(creatoropts["outdir"]) if creatoropts['release'] is not None: creator.release_output(ksconf, creatoropts['outdir'], creatoropts['release']) creator.print_outimage_info() except errors.CreatorError: raise finally: creator.cleanup() msger.info("Finished.") return 0 @classmethod def do_chroot(cls, target, cmd=[]): os_image = cls.do_unpack(target) os_image_dir = os.path.dirname(os_image) # unpack image to target dir imgsize = misc.get_file_size(os_image) * 1024L * 1024L imgtype = misc.get_image_type(os_image) if imgtype == "btrfsimg": fstype = "btrfs" myDiskMount = fs_related.BtrfsDiskMount elif imgtype in ("ext3fsimg", "ext4fsimg"): fstype = imgtype[:4] myDiskMount = fs_related.ExtDiskMount else: raise errors.CreatorError("Unsupported filesystem type: %s" % fstype) extmnt = misc.mkdtemp() extloop = myDiskMount(fs_related.SparseLoopbackDisk(os_image, imgsize), extmnt, fstype, 4096, "%s label" % fstype) try: extloop.mount() except errors.MountError: extloop.cleanup() shutil.rmtree(extmnt, ignore_errors = True) shutil.rmtree(os_image_dir, ignore_errors = True) raise try: if len(cmd) != 0: cmdline = ' '.join(cmd) else: cmdline = "/bin/bash" envcmd = fs_related.find_binary_inchroot("env", extmnt) if envcmd: cmdline = "%s HOME=/root %s" % (envcmd, cmdline) chroot.chroot(extmnt, None, cmdline) except: raise errors.CreatorError("Failed to chroot to %s." %target) finally: chroot.cleanup_after_chroot("img", extloop, os_image_dir, extmnt) @classmethod def do_pack(cls, base_on): import subprocess def __mkinitrd(instance): kernelver = instance._get_kernel_versions().values()[0][0] args = [ "/usr/libexec/mkliveinitrd", "/boot/initrd-%s.img" % kernelver, "%s" % kernelver ] try: subprocess.call(args, preexec_fn = instance._chroot) except OSError, (err, msg): raise errors.CreatorError("Failed to execute /usr/libexec/mkliveinitrd: %s" % msg) def __run_post_cleanups(instance): kernelver = instance._get_kernel_versions().values()[0][0] args = ["rm", "-f", "/boot/initrd-%s.img" % kernelver] try: subprocess.call(args, preexec_fn = instance._chroot) except OSError, (err, msg): raise errors.CreatorError("Failed to run post cleanups: %s" % msg) convertoropts = configmgr.convert convertoropts['name'] = os.path.splitext(os.path.basename(base_on))[0] convertor = livecd.LiveCDImageCreator(convertoropts) imgtype = misc.get_image_type(base_on) if imgtype == "btrfsimg": fstype = "btrfs" elif imgtype in ("ext3fsimg", "ext4fsimg"): fstype = imgtype[:4] else: raise errors.CreatorError("Unsupported filesystem type: %s" % fstype) convertor._set_fstype(fstype) try: convertor.mount(base_on) __mkinitrd(convertor) convertor._create_bootconfig() __run_post_cleanups(convertor) convertor.launch_shell(convertoropts['shell']) convertor.unmount() convertor.package() convertor.print_outimage_info() finally: shutil.rmtree(os.path.dirname(base_on), ignore_errors = True) @classmethod def do_unpack(cls, srcimg): img = srcimg imgmnt = misc.mkdtemp() imgloop = fs_related.DiskMount(fs_related.LoopbackDisk(img, 0), imgmnt) try: imgloop.mount() except errors.MountError: imgloop.cleanup() raise # legacy LiveOS filesystem layout support, remove for F9 or F10 if os.path.exists(imgmnt + "/squashfs.img"): squashimg = imgmnt + "/squashfs.img" else: squashimg = imgmnt + "/LiveOS/squashfs.img" tmpoutdir = misc.mkdtemp() # unsquashfs requires outdir mustn't exist shutil.rmtree(tmpoutdir, ignore_errors = True) misc.uncompress_squashfs(squashimg, tmpoutdir) try: # legacy LiveOS filesystem layout support, remove for F9 or F10 if os.path.exists(tmpoutdir + "/os.img"): os_image = tmpoutdir + "/os.img" else: os_image = tmpoutdir + "/LiveOS/ext3fs.img" if not os.path.exists(os_image): raise errors.CreatorError("'%s' is not a valid live CD ISO : neither " "LiveOS/ext3fs.img nor os.img exist" %img) imgname = os.path.basename(srcimg) imgname = os.path.splitext(imgname)[0] + ".img" rtimage = os.path.join(tempfile.mkdtemp(dir = "/var/tmp", prefix = "tmp"), imgname) shutil.copyfile(os_image, rtimage) finally: imgloop.cleanup() shutil.rmtree(tmpoutdir, ignore_errors = True) shutil.rmtree(imgmnt, ignore_errors = True) return rtimage ########NEW FILE######## __FILENAME__ = liveusb_plugin #!/usr/bin/python -tt # # Copyright (c) 2011 Intel, Inc. # # This program is free software; you can redistribute it and/or modify it # under the terms of the GNU General Public License as published by the Free # Software Foundation; version 2 of the License # # This program is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY # or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License # for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., 59 # Temple Place - Suite 330, Boston, MA 02111-1307, USA. import os import shutil import tempfile from mic import chroot, msger from mic.utils import misc, fs_related, errors from mic.utils.partitionedfs import PartitionedMount from mic.conf import configmgr from mic.plugin import pluginmgr import mic.imager.liveusb as liveusb from mic.pluginbase import ImagerPlugin class LiveUSBPlugin(ImagerPlugin): name = 'liveusb' @classmethod def do_create(self, subcmd, opts, *args): """${cmd_name}: create liveusb image Usage: ${name} ${cmd_name} <ksfile> [OPTS] ${cmd_option_list} """ if len(args) != 1: raise errors.Usage("Extra arguments given") creatoropts = configmgr.create ksconf = args[0] if creatoropts['runtime'] == "bootstrap": configmgr._ksconf = ksconf rt_util.bootstrap_mic() if creatoropts['arch'] and creatoropts['arch'].startswith('arm'): msger.warning('liveusb cannot support arm images, Quit') return recording_pkgs = [] if len(creatoropts['record_pkgs']) > 0: recording_pkgs = creatoropts['record_pkgs'] if creatoropts['release'] is not None: if 'name' not in recording_pkgs: recording_pkgs.append('name') if 'vcs' not in recording_pkgs: recording_pkgs.append('vcs') configmgr._ksconf = ksconf # Called After setting the configmgr._ksconf as the creatoropts['name'] is reset there. if creatoropts['release'] is not None: creatoropts['outdir'] = "%s/%s/images/%s/" % (creatoropts['outdir'], creatoropts['release'], creatoropts['name']) # try to find the pkgmgr pkgmgr = None backends = pluginmgr.get_plugins('backend') if 'auto' == creatoropts['pkgmgr']: for key in configmgr.prefer_backends: if key in backends: pkgmgr = backends[key] break else: for key in backends.keys(): if key == creatoropts['pkgmgr']: pkgmgr = backends[key] break if not pkgmgr: raise errors.CreatorError("Can't find backend: %s, " "available choices: %s" % (creatoropts['pkgmgr'], ','.join(backends.keys()))) creator = liveusb.LiveUSBImageCreator(creatoropts, pkgmgr) if len(recording_pkgs) > 0: creator._recording_pkgs = recording_pkgs self.check_image_exists(creator.destdir, creator.pack_to, [creator.name + ".usbimg"], creatoropts['release']) try: creator.check_depend_tools() creator.mount(None, creatoropts["cachedir"]) creator.install() creator.configure(creatoropts["repomd"]) creator.copy_kernel() creator.unmount() creator.package(creatoropts["outdir"]) if creatoropts['release'] is not None: creator.release_output(ksconf, creatoropts['outdir'], creatoropts['release']) creator.print_outimage_info() except errors.CreatorError: raise finally: creator.cleanup() msger.info("Finished.") return 0 @classmethod def do_chroot(cls, target, cmd=[]): os_image = cls.do_unpack(target) os_image_dir = os.path.dirname(os_image) # unpack image to target dir imgsize = misc.get_file_size(os_image) * 1024L * 1024L imgtype = misc.get_image_type(os_image) if imgtype == "btrfsimg": fstype = "btrfs" myDiskMount = fs_related.BtrfsDiskMount elif imgtype in ("ext3fsimg", "ext4fsimg"): fstype = imgtype[:4] myDiskMount = fs_related.ExtDiskMount else: raise errors.CreatorError("Unsupported filesystem type: %s" % fstype) extmnt = misc.mkdtemp() extloop = myDiskMount(fs_related.SparseLoopbackDisk(os_image, imgsize), extmnt, fstype, 4096, "%s label" % fstype) try: extloop.mount() except errors.MountError: extloop.cleanup() shutil.rmtree(extmnt, ignore_errors = True) raise try: if len(cmd) != 0: cmdline = ' '.join(cmd) else: cmdline = "/bin/bash" envcmd = fs_related.find_binary_inchroot("env", extmnt) if envcmd: cmdline = "%s HOME=/root %s" % (envcmd, cmdline) chroot.chroot(extmnt, None, cmdline) except: raise errors.CreatorError("Failed to chroot to %s." %target) finally: chroot.cleanup_after_chroot("img", extloop, os_image_dir, extmnt) @classmethod def do_pack(cls, base_on): import subprocess def __mkinitrd(instance): kernelver = instance._get_kernel_versions().values()[0][0] args = [ "/usr/libexec/mkliveinitrd", "/boot/initrd-%s.img" % kernelver, "%s" % kernelver ] try: subprocess.call(args, preexec_fn = instance._chroot) except OSError, (err, msg): raise errors.CreatorError("Failed to execute /usr/libexec/mkliveinitrd: %s" % msg) def __run_post_cleanups(instance): kernelver = instance._get_kernel_versions().values()[0][0] args = ["rm", "-f", "/boot/initrd-%s.img" % kernelver] try: subprocess.call(args, preexec_fn = instance._chroot) except OSError, (err, msg): raise errors.CreatorError("Failed to run post cleanups: %s" % msg) convertoropts = configmgr.convert convertoropts['name'] = os.path.splitext(os.path.basename(base_on))[0] convertor = liveusb.LiveUSBImageCreator(convertoropts) imgtype = misc.get_image_type(base_on) if imgtype == "btrfsimg": fstype = "btrfs" elif imgtype in ("ext3fsimg", "ext4fsimg"): fstype = imgtype[:4] else: raise errors.CreatorError("Unsupported filesystem type: %s" % fstyp) convertor._set_fstype(fstype) try: convertor.mount(base_on) __mkinitrd(convertor) convertor._create_bootconfig() __run_post_cleanups(convertor) convertor.launch_shell(convertoropts['shell']) convertor.unmount() convertor.package() convertor.print_outimage_info() finally: shutil.rmtree(os.path.dirname(base_on), ignore_errors = True) @classmethod def do_unpack(cls, srcimg): img = srcimg imgsize = misc.get_file_size(img) * 1024L * 1024L imgmnt = misc.mkdtemp() disk = fs_related.SparseLoopbackDisk(img, imgsize) imgloop = PartitionedMount(imgmnt, skipformat = True) imgloop.add_disk('/dev/sdb', disk) imgloop.add_partition(imgsize/1024/1024, "/dev/sdb", "/", "vfat", boot=False) try: imgloop.mount() except errors.MountError: imgloop.cleanup() raise # legacy LiveOS filesystem layout support, remove for F9 or F10 if os.path.exists(imgmnt + "/squashfs.img"): squashimg = imgmnt + "/squashfs.img" else: squashimg = imgmnt + "/LiveOS/squashfs.img" tmpoutdir = misc.mkdtemp() # unsquashfs requires outdir mustn't exist shutil.rmtree(tmpoutdir, ignore_errors = True) misc.uncompress_squashfs(squashimg, tmpoutdir) try: # legacy LiveOS filesystem layout support, remove for F9 or F10 if os.path.exists(tmpoutdir + "/os.img"): os_image = tmpoutdir + "/os.img" else: os_image = tmpoutdir + "/LiveOS/ext3fs.img" if not os.path.exists(os_image): raise errors.CreatorError("'%s' is not a valid live CD ISO : neither " "LiveOS/ext3fs.img nor os.img exist" %img) imgname = os.path.basename(srcimg) imgname = os.path.splitext(imgname)[0] + ".img" rtimage = os.path.join(tempfile.mkdtemp(dir = "/var/tmp", prefix = "tmp"), imgname) shutil.copyfile(os_image, rtimage) finally: imgloop.cleanup() shutil.rmtree(tmpoutdir, ignore_errors = True) shutil.rmtree(imgmnt, ignore_errors = True) return rtimage ########NEW FILE######## __FILENAME__ = loop_plugin #!/usr/bin/python -tt # # Copyright (c) 2011 Intel, Inc. # # This program is free software; you can redistribute it and/or modify it # under the terms of the GNU General Public License as published by the Free # Software Foundation; version 2 of the License # # This program is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY # or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License # for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., 59 # Temple Place - Suite 330, Boston, MA 02111-1307, USA. import os import shutil import tempfile from mic import chroot, msger from mic.utils import misc, fs_related, errors, cmdln from mic.conf import configmgr from mic.plugin import pluginmgr from mic.imager.loop import LoopImageCreator, load_mountpoints from mic.pluginbase import ImagerPlugin class LoopPlugin(ImagerPlugin): name = 'loop' @classmethod @cmdln.option("--compress-disk-image", dest="compress_image", type='choice', choices=("gz", "bz2"), default=None, help="Same with --compress-image") # alias to compress-image for compatibility @cmdln.option("--compress-image", dest="compress_image", type='choice', choices=("gz", "bz2"), default=None, help="Compress all loop images with 'gz' or 'bz2'") @cmdln.option("--shrink", action='store_true', default=False, help="Whether to shrink loop images to minimal size") def do_create(self, subcmd, opts, *args): """${cmd_name}: create loop image Usage: ${name} ${cmd_name} <ksfile> [OPTS] ${cmd_option_list} """ if len(args) != 1: raise errors.Usage("Extra arguments given") creatoropts = configmgr.create ksconf = args[0] if creatoropts['runtime'] == "bootstrap": configmgr._ksconf = ksconf rt_util.bootstrap_mic() recording_pkgs = [] if len(creatoropts['record_pkgs']) > 0: recording_pkgs = creatoropts['record_pkgs'] if creatoropts['release'] is not None: if 'name' not in recording_pkgs: recording_pkgs.append('name') if 'vcs' not in recording_pkgs: recording_pkgs.append('vcs') configmgr._ksconf = ksconf # Called After setting the configmgr._ksconf # as the creatoropts['name'] is reset there. if creatoropts['release'] is not None: creatoropts['outdir'] = "%s/%s/images/%s/" % (creatoropts['outdir'], creatoropts['release'], creatoropts['name']) # try to find the pkgmgr pkgmgr = None backends = pluginmgr.get_plugins('backend') if 'auto' == creatoropts['pkgmgr']: for key in configmgr.prefer_backends: if key in backends: pkgmgr = backends[key] break else: for key in backends.keys(): if key == creatoropts['pkgmgr']: pkgmgr = backends[key] break if not pkgmgr: raise errors.CreatorError("Can't find backend: %s, " "available choices: %s" % (creatoropts['pkgmgr'], ','.join(backends.keys()))) creator = LoopImageCreator(creatoropts, pkgmgr, opts.compress_image, opts.shrink) if len(recording_pkgs) > 0: creator._recording_pkgs = recording_pkgs image_names = [creator.name + ".img"] image_names.extend(creator.get_image_names()) self.check_image_exists(creator.destdir, creator.pack_to, image_names, creatoropts['release']) try: creator.check_depend_tools() creator.mount(None, creatoropts["cachedir"]) creator.install() creator.configure(creatoropts["repomd"]) creator.copy_kernel() creator.unmount() creator.package(creatoropts["outdir"]) if creatoropts['release'] is not None: creator.release_output(ksconf, creatoropts['outdir'], creatoropts['release']) creator.print_outimage_info() except errors.CreatorError: raise finally: creator.cleanup() msger.info("Finished.") return 0 @classmethod def _do_chroot_tar(cls, target, cmd=[]): mountfp_xml = os.path.splitext(target)[0] + '.xml' if not os.path.exists(mountfp_xml): raise errors.CreatorError("No mount point file found for this tar " "image, please check %s" % mountfp_xml) import tarfile tar = tarfile.open(target, 'r') tmpdir = misc.mkdtemp() tar.extractall(path=tmpdir) tar.close() mntdir = misc.mkdtemp() loops = [] for (mp, label, name, size, fstype) in load_mountpoints(mountfp_xml): if fstype in ("ext2", "ext3", "ext4"): myDiskMount = fs_related.ExtDiskMount elif fstype == "btrfs": myDiskMount = fs_related.BtrfsDiskMount elif fstype in ("vfat", "msdos"): myDiskMount = fs_related.VfatDiskMount else: msger.error("Cannot support fstype: %s" % fstype) name = os.path.join(tmpdir, name) size = size * 1024L * 1024L loop = myDiskMount(fs_related.SparseLoopbackDisk(name, size), os.path.join(mntdir, mp.lstrip('/')), fstype, size, label) try: msger.verbose("Mount %s to %s" % (mp, mntdir + mp)) fs_related.makedirs(os.path.join(mntdir, mp.lstrip('/'))) loop.mount() except: loop.cleanup() for lp in reversed(loops): chroot.cleanup_after_chroot("img", lp, None, mntdir) shutil.rmtree(tmpdir, ignore_errors=True) raise loops.append(loop) try: if len(cmd) != 0: cmdline = "/usr/bin/env HOME=/root " + ' '.join(cmd) else: cmdline = "/usr/bin/env HOME=/root /bin/bash" chroot.chroot(mntdir, None, cmdline) except: raise errors.CreatorError("Failed to chroot to %s." % target) finally: for loop in reversed(loops): chroot.cleanup_after_chroot("img", loop, None, mntdir) shutil.rmtree(tmpdir, ignore_errors=True) @classmethod def do_chroot(cls, target, cmd=[]): if target.endswith('.tar'): import tarfile if tarfile.is_tarfile(target): LoopPlugin._do_chroot_tar(target, cmd) return else: raise errors.CreatorError("damaged tarball for loop images") img = target imgsize = misc.get_file_size(img) * 1024L * 1024L imgtype = misc.get_image_type(img) if imgtype == "btrfsimg": fstype = "btrfs" myDiskMount = fs_related.BtrfsDiskMount elif imgtype in ("ext3fsimg", "ext4fsimg"): fstype = imgtype[:4] myDiskMount = fs_related.ExtDiskMount else: raise errors.CreatorError("Unsupported filesystem type: %s" \ % imgtype) extmnt = misc.mkdtemp() extloop = myDiskMount(fs_related.SparseLoopbackDisk(img, imgsize), extmnt, fstype, 4096, "%s label" % fstype) try: extloop.mount() except errors.MountError: extloop.cleanup() shutil.rmtree(extmnt, ignore_errors=True) raise try: if len(cmd) != 0: cmdline = ' '.join(cmd) else: cmdline = "/bin/bash" envcmd = fs_related.find_binary_inchroot("env", extmnt) if envcmd: cmdline = "%s HOME=/root %s" % (envcmd, cmdline) chroot.chroot(extmnt, None, cmdline) except: raise errors.CreatorError("Failed to chroot to %s." % img) finally: chroot.cleanup_after_chroot("img", extloop, None, extmnt) @classmethod def do_unpack(cls, srcimg): image = os.path.join(tempfile.mkdtemp(dir="/var/tmp", prefix="tmp"), "target.img") msger.info("Copying file system ...") shutil.copyfile(srcimg, image) return image ########NEW FILE######## __FILENAME__ = raw_plugin #!/usr/bin/python -tt # # Copyright (c) 2011 Intel, Inc. # # This program is free software; you can redistribute it and/or modify it # under the terms of the GNU General Public License as published by the Free # Software Foundation; version 2 of the License # # This program is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY # or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License # for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., 59 # Temple Place - Suite 330, Boston, MA 02111-1307, USA. import os import shutil import re import tempfile from mic import chroot, msger from mic.utils import misc, fs_related, errors, runner, cmdln from mic.conf import configmgr from mic.plugin import pluginmgr from mic.utils.partitionedfs import PartitionedMount import mic.imager.raw as raw from mic.pluginbase import ImagerPlugin class RawPlugin(ImagerPlugin): name = 'raw' @classmethod @cmdln.option("--compress-disk-image", dest="compress_image", type='choice', choices=("gz", "bz2"), default=None, help="Same with --compress-image") @cmdln.option("--compress-image", dest="compress_image", type='choice', choices=("gz", "bz2"), default = None, help="Compress all raw images before package") @cmdln.option("--generate-bmap", action="store_true", default = None, help="also generate the block map file") @cmdln.option("--fstab-entry", dest="fstab_entry", type='choice', choices=("name", "uuid"), default="uuid", help="Set fstab entry, 'name' means using device names, " "'uuid' means using filesystem uuid") def do_create(self, subcmd, opts, *args): """${cmd_name}: create raw image Usage: ${name} ${cmd_name} <ksfile> [OPTS] ${cmd_option_list} """ if len(args) != 1: raise errors.Usage("Extra arguments given") creatoropts = configmgr.create ksconf = args[0] if creatoropts['runtime'] == "bootstrap": configmgr._ksconf = ksconf rt_util.bootstrap_mic() recording_pkgs = [] if len(creatoropts['record_pkgs']) > 0: recording_pkgs = creatoropts['record_pkgs'] if creatoropts['release'] is not None: if 'name' not in recording_pkgs: recording_pkgs.append('name') if 'vcs' not in recording_pkgs: recording_pkgs.append('vcs') configmgr._ksconf = ksconf # Called After setting the configmgr._ksconf as the creatoropts['name'] is reset there. if creatoropts['release'] is not None: creatoropts['outdir'] = "%s/%s/images/%s/" % (creatoropts['outdir'], creatoropts['release'], creatoropts['name']) # try to find the pkgmgr pkgmgr = None backends = pluginmgr.get_plugins('backend') if 'auto' == creatoropts['pkgmgr']: for key in configmgr.prefer_backends: if key in backends: pkgmgr = backends[key] break else: for key in backends.keys(): if key == creatoropts['pkgmgr']: pkgmgr = backends[key] break if not pkgmgr: raise errors.CreatorError("Can't find backend: %s, " "available choices: %s" % (creatoropts['pkgmgr'], ','.join(backends.keys()))) creator = raw.RawImageCreator(creatoropts, pkgmgr, opts.compress_image, opts.generate_bmap, opts.fstab_entry) if len(recording_pkgs) > 0: creator._recording_pkgs = recording_pkgs images = ["%s-%s.raw" % (creator.name, disk_name) for disk_name in creator.get_disk_names()] self.check_image_exists(creator.destdir, creator.pack_to, images, creatoropts['release']) try: creator.check_depend_tools() creator.mount(None, creatoropts["cachedir"]) creator.install() creator.configure(creatoropts["repomd"]) creator.copy_kernel() creator.unmount() creator.generate_bmap() creator.package(creatoropts["outdir"]) if creatoropts['release'] is not None: creator.release_output(ksconf, creatoropts['outdir'], creatoropts['release']) creator.print_outimage_info() except errors.CreatorError: raise finally: creator.cleanup() msger.info("Finished.") return 0 @classmethod def do_chroot(cls, target, cmd=[]): img = target imgsize = misc.get_file_size(img) * 1024L * 1024L partedcmd = fs_related.find_binary_path("parted") disk = fs_related.SparseLoopbackDisk(img, imgsize) imgmnt = misc.mkdtemp() imgloop = PartitionedMount(imgmnt, skipformat = True) imgloop.add_disk('/dev/sdb', disk) img_fstype = "ext3" msger.info("Partition Table:") partnum = [] for line in runner.outs([partedcmd, "-s", img, "print"]).splitlines(): # no use strip to keep line output here if "Number" in line: msger.raw(line) if line.strip() and line.strip()[0].isdigit(): partnum.append(line.strip()[0]) msger.raw(line) rootpart = None if len(partnum) > 1: rootpart = msger.choice("please choose root partition", partnum) # Check the partitions from raw disk. # if choose root part, the mark it as mounted if rootpart: root_mounted = True else: root_mounted = False partition_mounts = 0 for line in runner.outs([partedcmd,"-s",img,"unit","B","print"]).splitlines(): line = line.strip() # Lines that start with number are the partitions, # because parted can be translated we can't refer to any text lines. if not line or not line[0].isdigit(): continue # Some vars have extra , as list seperator. line = line.replace(",","") # Example of parted output lines that are handled: # Number Start End Size Type File system Flags # 1 512B 3400000511B 3400000000B primary # 2 3400531968B 3656384511B 255852544B primary linux-swap(v1) # 3 3656384512B 3720347647B 63963136B primary fat16 boot, lba partition_info = re.split("\s+",line) size = partition_info[3].split("B")[0] if len(partition_info) < 6 or partition_info[5] in ["boot"]: # No filesystem can be found from partition line. Assuming # btrfs, because that is the only MeeGo fs that parted does # not recognize properly. # TODO: Can we make better assumption? fstype = "btrfs" elif partition_info[5] in ["ext2","ext3","ext4","btrfs"]: fstype = partition_info[5] elif partition_info[5] in ["fat16","fat32"]: fstype = "vfat" elif "swap" in partition_info[5]: fstype = "swap" else: raise errors.CreatorError("Could not recognize partition fs type '%s'." % partition_info[5]) if rootpart and rootpart == line[0]: mountpoint = '/' elif not root_mounted and fstype in ["ext2","ext3","ext4","btrfs"]: # TODO: Check that this is actually the valid root partition from /etc/fstab mountpoint = "/" root_mounted = True elif fstype == "swap": mountpoint = "swap" else: # TODO: Assing better mount points for the rest of the partitions. partition_mounts += 1 mountpoint = "/media/partition_%d" % partition_mounts if "boot" in partition_info: boot = True else: boot = False msger.verbose("Size: %s Bytes, fstype: %s, mountpoint: %s, boot: %s" % (size, fstype, mountpoint, boot)) # TODO: add_partition should take bytes as size parameter. imgloop.add_partition((int)(size)/1024/1024, "/dev/sdb", mountpoint, fstype = fstype, boot = boot) try: imgloop.mount() except errors.MountError: imgloop.cleanup() raise try: if len(cmd) != 0: cmdline = ' '.join(cmd) else: cmdline = "/bin/bash" envcmd = fs_related.find_binary_inchroot("env", imgmnt) if envcmd: cmdline = "%s HOME=/root %s" % (envcmd, cmdline) chroot.chroot(imgmnt, None, cmdline) except: raise errors.CreatorError("Failed to chroot to %s." %img) finally: chroot.cleanup_after_chroot("img", imgloop, None, imgmnt) @classmethod def do_unpack(cls, srcimg): srcimgsize = (misc.get_file_size(srcimg)) * 1024L * 1024L srcmnt = misc.mkdtemp("srcmnt") disk = fs_related.SparseLoopbackDisk(srcimg, srcimgsize) srcloop = PartitionedMount(srcmnt, skipformat = True) srcloop.add_disk('/dev/sdb', disk) srcloop.add_partition(srcimgsize/1024/1024, "/dev/sdb", "/", "ext3", boot=False) try: srcloop.mount() except errors.MountError: srcloop.cleanup() raise image = os.path.join(tempfile.mkdtemp(dir = "/var/tmp", prefix = "tmp"), "target.img") args = ['dd', "if=%s" % srcloop.partitions[0]['device'], "of=%s" % image] msger.info("`dd` image ...") rc = runner.show(args) srcloop.cleanup() shutil.rmtree(os.path.dirname(srcmnt), ignore_errors = True) if rc != 0: raise errors.CreatorError("Failed to dd") else: return image ########NEW FILE######## __FILENAME__ = bootimg-efi # ex:ts=4:sw=4:sts=4:et # -*- tab-width: 4; c-basic-offset: 4; indent-tabs-mode: nil -*- # # Copyright (c) 2014, Intel Corporation. # All rights reserved. # # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License version 2 as # published by the Free Software Foundation. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., # 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA. # # DESCRIPTION # This implements the 'bootimg-efi' source plugin class for 'wic' # # AUTHORS # Tom Zanussi <tom.zanussi (at] linux.intel.com> # import os import shutil import re import tempfile from mic import kickstart, chroot, msger from mic.utils import misc, fs_related, errors, runner, cmdln from mic.conf import configmgr from mic.plugin import pluginmgr from mic.utils.partitionedfs import PartitionedMount import mic.imager.direct as direct from mic.pluginbase import SourcePlugin from mic.utils.oe.misc import * from mic.imager.direct import DirectImageCreator class BootimgEFIPlugin(SourcePlugin): name = 'bootimg-efi' @classmethod def do_configure_partition(self, part, cr, cr_workdir, oe_builddir, bootimg_dir, kernel_dir, native_sysroot): """ Called before do_prepare_partition(), creates grubefi config """ hdddir = "%s/hdd/boot" % cr_workdir rm_cmd = "rm -rf %s" % cr_workdir exec_cmd(rm_cmd) install_cmd = "install -d %s/EFI/BOOT" % hdddir tmp = exec_cmd(install_cmd) splash = os.path.join(cr_workdir, "/EFI/boot/splash.jpg") if os.path.exists(splash): splashline = "menu background splash.jpg" else: splashline = "" (rootdev, root_part_uuid) = cr._get_boot_config() options = cr.ks.handler.bootloader.appendLine grubefi_conf = "" grubefi_conf += "serial --unit=0 --speed=115200 --word=8 --parity=no --stop=1\n" grubefi_conf += "default=boot\n" timeout = kickstart.get_timeout(cr.ks) if not timeout: timeout = 0 grubefi_conf += "timeout=%s\n" % timeout grubefi_conf += "menuentry 'boot'{\n" kernel = "/vmlinuz" if cr._ptable_format == 'msdos': rootstr = rootdev else: if not root_part_uuid: raise MountError("Cannot find the root GPT partition UUID") rootstr = "PARTUUID=%s" % root_part_uuid grubefi_conf += "linux %s root=%s rootwait %s\n" \ % (kernel, rootstr, options) grubefi_conf += "}\n" if splashline: syslinux_conf += "%s\n" % splashline msger.debug("Writing grubefi config %s/hdd/boot/EFI/BOOT/grub.cfg" \ % cr_workdir) cfg = open("%s/hdd/boot/EFI/BOOT/grub.cfg" % cr_workdir, "w") cfg.write(grubefi_conf) cfg.close() @classmethod def do_prepare_partition(self, part, cr, cr_workdir, oe_builddir, bootimg_dir, kernel_dir, rootfs_dir, native_sysroot): """ Called to do the actual content population for a partition i.e. it 'prepares' the partition to be incorporated into the image. In this case, prepare content for an EFI (grub) boot partition. """ if not bootimg_dir: bootimg_dir = get_bitbake_var("HDDDIR") if not bootimg_dir: msger.error("Couldn't find HDDDIR, exiting\n") # just so the result notes display it cr.set_bootimg_dir(bootimg_dir) staging_kernel_dir = kernel_dir staging_data_dir = bootimg_dir hdddir = "%s/hdd" % cr_workdir install_cmd = "install -m 0644 %s/bzImage %s/bzImage" % \ (staging_kernel_dir, hdddir) tmp = exec_cmd(install_cmd) shutil.copyfile("%s/hdd/boot/EFI/BOOT/grub.cfg" % cr_workdir, "%s/grub.cfg" % cr_workdir) cp_cmd = "cp %s/EFI/BOOT/* %s/EFI/BOOT" % (staging_data_dir, hdddir) exec_cmd(cp_cmd, True) shutil.move("%s/grub.cfg" % cr_workdir, "%s/hdd/boot/EFI/BOOT/grub.cfg" % cr_workdir) du_cmd = "du -bks %s" % hdddir rc, out = exec_cmd(du_cmd) blocks = int(out.split()[0]) extra_blocks = part.get_extra_block_count(blocks) if extra_blocks < BOOTDD_EXTRA_SPACE: extra_blocks = BOOTDD_EXTRA_SPACE blocks += extra_blocks msger.debug("Added %d extra blocks to %s to get to %d total blocks" % \ (extra_blocks, part.mountpoint, blocks)) # Ensure total sectors is an integral number of sectors per # track or mcopy will complain. Sectors are 512 bytes, and we # generate images with 32 sectors per track. This calculation is # done in blocks, thus the mod by 16 instead of 32. blocks += (16 - (blocks % 16)) # dosfs image, created by mkdosfs bootimg = "%s/boot.img" % cr_workdir dosfs_cmd = "mkdosfs -n efi -C %s %d" % (bootimg, blocks) exec_native_cmd(dosfs_cmd, native_sysroot) mcopy_cmd = "mcopy -i %s -s %s/* ::/" % (bootimg, hdddir) exec_native_cmd(mcopy_cmd, native_sysroot) chmod_cmd = "chmod 644 %s" % bootimg exec_cmd(chmod_cmd) du_cmd = "du -Lbms %s" % bootimg rc, out = exec_cmd(du_cmd) bootimg_size = out.split()[0] part.set_size(bootimg_size) part.set_source_file(bootimg) ########NEW FILE######## __FILENAME__ = bootimg-pcbios # ex:ts=4:sw=4:sts=4:et # -*- tab-width: 4; c-basic-offset: 4; indent-tabs-mode: nil -*- # # Copyright (c) 2014, Intel Corporation. # All rights reserved. # # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License version 2 as # published by the Free Software Foundation. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., # 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA. # # DESCRIPTION # This implements the 'bootimg-pcbios' source plugin class for 'wic' # # AUTHORS # Tom Zanussi <tom.zanussi (at] linux.intel.com> # import os import shutil import re import tempfile from mic import kickstart, chroot, msger from mic.utils import misc, fs_related, errors, runner, cmdln from mic.conf import configmgr from mic.plugin import pluginmgr from mic.utils.partitionedfs import PartitionedMount import mic.imager.direct as direct from mic.pluginbase import SourcePlugin from mic.utils.oe.misc import * from mic.imager.direct import DirectImageCreator class BootimgPcbiosPlugin(SourcePlugin): name = 'bootimg-pcbios' @classmethod def do_install_disk(self, disk, disk_name, cr, workdir, oe_builddir, bootimg_dir, kernel_dir, native_sysroot): """ Called after all partitions have been prepared and assembled into a disk image. In this case, we install the MBR. """ mbrfile = "%s/syslinux/" % bootimg_dir if cr._ptable_format == 'gpt': mbrfile += "gptmbr.bin" else: mbrfile += "mbr.bin" if not os.path.exists(mbrfile): msger.error("Couldn't find %s. If using the -e option, do you have the right MACHINE set in local.conf? If not, is the bootimg_dir path correct?" % mbrfile) full_path = cr._full_path(workdir, disk_name, "direct") msger.debug("Installing MBR on disk %s as %s with size %s bytes" \ % (disk_name, full_path, disk['min_size'])) rc = runner.show(['dd', 'if=%s' % mbrfile, 'of=%s' % full_path, 'conv=notrunc']) if rc != 0: raise MountError("Unable to set MBR to %s" % full_path) @classmethod def do_configure_partition(self, part, cr, cr_workdir, oe_builddir, bootimg_dir, kernel_dir, native_sysroot): """ Called before do_prepare_partition(), creates syslinux config """ hdddir = "%s/hdd/boot" % cr_workdir rm_cmd = "rm -rf " + cr_workdir exec_cmd(rm_cmd) install_cmd = "install -d %s" % hdddir tmp = exec_cmd(install_cmd) splash = os.path.join(cr_workdir, "/hdd/boot/splash.jpg") if os.path.exists(splash): splashline = "menu background splash.jpg" else: splashline = "" (rootdev, root_part_uuid) = cr._get_boot_config() options = cr.ks.handler.bootloader.appendLine syslinux_conf = "" syslinux_conf += "PROMPT 0\n" timeout = kickstart.get_timeout(cr.ks) if not timeout: timeout = 0 syslinux_conf += "TIMEOUT " + str(timeout) + "\n" syslinux_conf += "\n" syslinux_conf += "ALLOWOPTIONS 1\n" syslinux_conf += "SERIAL 0 115200\n" syslinux_conf += "\n" if splashline: syslinux_conf += "%s\n" % splashline syslinux_conf += "DEFAULT boot\n" syslinux_conf += "LABEL boot\n" kernel = "/vmlinuz" syslinux_conf += "KERNEL " + kernel + "\n" if cr._ptable_format == 'msdos': rootstr = rootdev else: if not root_part_uuid: raise MountError("Cannot find the root GPT partition UUID") rootstr = "PARTUUID=%s" % root_part_uuid syslinux_conf += "APPEND label=boot root=%s %s\n" % (rootstr, options) msger.debug("Writing syslinux config %s/hdd/boot/syslinux.cfg" \ % cr_workdir) cfg = open("%s/hdd/boot/syslinux.cfg" % cr_workdir, "w") cfg.write(syslinux_conf) cfg.close() @classmethod def do_prepare_partition(self, part, cr, cr_workdir, oe_builddir, bootimg_dir, kernel_dir, rootfs_dir, native_sysroot): """ Called to do the actual content population for a partition i.e. it 'prepares' the partition to be incorporated into the image. In this case, prepare content for legacy bios boot partition. """ if not bootimg_dir: bootimg_dir = get_bitbake_var("STAGING_DATADIR") if not bootimg_dir: msger.error("Couldn't find STAGING_DATADIR, exiting\n") # just so the result notes display it cr.set_bootimg_dir(bootimg_dir) staging_kernel_dir = kernel_dir staging_data_dir = bootimg_dir hdddir = "%s/hdd/boot" % cr_workdir install_cmd = "install -m 0644 %s/bzImage %s/vmlinuz" \ % (staging_kernel_dir, hdddir) tmp = exec_cmd(install_cmd) install_cmd = "install -m 444 %s/syslinux/ldlinux.sys %s/ldlinux.sys" \ % (staging_data_dir, hdddir) tmp = exec_cmd(install_cmd) du_cmd = "du -bks %s" % hdddir rc, out = exec_cmd(du_cmd) blocks = int(out.split()[0]) extra_blocks = part.get_extra_block_count(blocks) if extra_blocks < BOOTDD_EXTRA_SPACE: extra_blocks = BOOTDD_EXTRA_SPACE blocks += extra_blocks msger.debug("Added %d extra blocks to %s to get to %d total blocks" % \ (extra_blocks, part.mountpoint, blocks)) # Ensure total sectors is an integral number of sectors per # track or mcopy will complain. Sectors are 512 bytes, and we # generate images with 32 sectors per track. This calculation is # done in blocks, thus the mod by 16 instead of 32. blocks += (16 - (blocks % 16)) # dosfs image, created by mkdosfs bootimg = "%s/boot.img" % cr_workdir dosfs_cmd = "mkdosfs -n boot -S 512 -C %s %d" % (bootimg, blocks) exec_native_cmd(dosfs_cmd, native_sysroot) mcopy_cmd = "mcopy -i %s -s %s/* ::/" % (bootimg, hdddir) exec_native_cmd(mcopy_cmd, native_sysroot) syslinux_cmd = "syslinux %s" % bootimg exec_native_cmd(syslinux_cmd, native_sysroot) chmod_cmd = "chmod 644 %s" % bootimg exec_cmd(chmod_cmd) du_cmd = "du -Lbms %s" % bootimg rc, out = exec_cmd(du_cmd) bootimg_size = out.split()[0] part.set_size(bootimg_size) part.set_source_file(bootimg) ########NEW FILE######## __FILENAME__ = rootfs # ex:ts=4:sw=4:sts=4:et # -*- tab-width: 4; c-basic-offset: 4; indent-tabs-mode: nil -*- # # Copyright (c) 2014, Intel Corporation. # All rights reserved. # # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License version 2 as # published by the Free Software Foundation. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., # 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA. # # DESCRIPTION # This implements the 'rootfs' source plugin class for 'wic' # # AUTHORS # Tom Zanussi <tom.zanussi (at] linux.intel.com> # Joao Henrique Ferreira de Freitas <joaohf (at] gmail.com> # import os import shutil import re import tempfile from mic import kickstart, chroot, msger from mic.utils import misc, fs_related, errors, runner, cmdln from mic.conf import configmgr from mic.plugin import pluginmgr from mic.utils.partitionedfs import PartitionedMount import mic.imager.direct as direct from mic.pluginbase import SourcePlugin from mic.utils.oe.misc import * from mic.imager.direct import DirectImageCreator class RootfsPlugin(SourcePlugin): name = 'rootfs' @staticmethod def __get_rootfs_dir(rootfs_dir): if os.path.isdir(rootfs_dir): return rootfs_dir bitbake_env_lines = find_bitbake_env_lines(rootfs_dir) if not bitbake_env_lines: msg = "Couldn't get bitbake environment, exiting." msger.error(msg) image_rootfs_dir = find_artifact(bitbake_env_lines, "IMAGE_ROOTFS") if not os.path.isdir(image_rootfs_dir): msg = "No valid artifact IMAGE_ROOTFS from image named" msg += " %s has been found at %s, exiting.\n" % \ (rootfs_dir, image_rootfs_dir) msger.error(msg) return image_rootfs_dir @classmethod def do_prepare_partition(self, part, cr, cr_workdir, oe_builddir, bootimg_dir, kernel_dir, krootfs_dir, native_sysroot): """ Called to do the actual content population for a partition i.e. it 'prepares' the partition to be incorporated into the image. In this case, prepare content for legacy bios boot partition. """ if part.rootfs is None: if not 'ROOTFS_DIR' in krootfs_dir: msg = "Couldn't find --rootfs-dir, exiting" msger.error(msg) rootfs_dir = krootfs_dir['ROOTFS_DIR'] else: if part.rootfs in krootfs_dir: rootfs_dir = krootfs_dir[part.rootfs] elif part.rootfs: rootfs_dir = part.rootfs else: msg = "Couldn't find --rootfs-dir=%s connection" msg += " or it is not a valid path, exiting" msger.error(msg % part.rootfs) real_rootfs_dir = self.__get_rootfs_dir(rootfs_dir) part.set_rootfs(real_rootfs_dir) part.prepare_rootfs(cr_workdir, oe_builddir, real_rootfs_dir, native_sysroot) ########NEW FILE######## __FILENAME__ = rt_util #!/usr/bin/python -tt # # Copyright (c) 2009, 2010, 2011 Intel, Inc. # # This program is free software; you can redistribute it and/or modify it # under the terms of the GNU General Public License as published by the Free # Software Foundation; version 2 of the License # # This program is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY # or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License # for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., 59 # Temple Place - Suite 330, Boston, MA 02111-1307, USA. from __future__ import with_statement import os import sys import glob import re import shutil import subprocess from mic import bootstrap, msger from mic.conf import configmgr from mic.utils import errors, proxy from mic.utils.fs_related import find_binary_path, makedirs from mic.chroot import setup_chrootenv, cleanup_chrootenv expath = lambda p: os.path.abspath(os.path.expanduser(p)) def bootstrap_mic(argv=None): def mychroot(): os.chroot(rootdir) os.chdir(cwd) # by default, sys.argv is used to run mic in bootstrap if not argv: argv = sys.argv if argv[0] not in ('/usr/bin/mic', 'mic'): argv[0] = '/usr/bin/mic' cropts = configmgr.create bsopts = configmgr.bootstrap distro = bsopts['distro_name'].lower() rootdir = bsopts['rootdir'] pkglist = bsopts['packages'] cwd = os.getcwd() # create bootstrap and run mic in bootstrap bsenv = bootstrap.Bootstrap(rootdir, distro, cropts['arch']) bsenv.logfile = cropts['logfile'] # rootdir is regenerated as a temp dir rootdir = bsenv.rootdir if 'optional' in bsopts: optlist = bsopts['optional'] else: optlist = [] try: msger.info("Creating %s bootstrap ..." % distro) bsenv.create(cropts['repomd'], pkglist, optlist) # bootstrap is relocated under "bootstrap" if os.path.exists(os.path.join(rootdir, "bootstrap")): rootdir = os.path.join(rootdir, "bootstrap") bsenv.dirsetup(rootdir) sync_mic(rootdir) #FIXME: sync the ks file to bootstrap if "/" == os.path.dirname(os.path.abspath(configmgr._ksconf)): safecopy(configmgr._ksconf, rootdir) msger.info("Start mic in bootstrap: %s\n" % rootdir) bindmounts = get_bindmounts(cropts) ret = bsenv.run(argv, cwd, rootdir, bindmounts) except errors.BootstrapError, err: msger.warning('\n%s' % err) if msger.ask("Switch to native mode and continue?"): return raise except RuntimeError, err: #change exception type but keep the trace back value, tb = sys.exc_info()[1:] raise errors.BootstrapError, value, tb else: sys.exit(ret) finally: bsenv.cleanup() def get_bindmounts(cropts): binddirs = [ os.getcwd(), cropts['tmpdir'], cropts['cachedir'], cropts['outdir'], cropts['local_pkgs_path'], ] bindfiles = [ cropts['logfile'], configmgr._ksconf, ] for lrepo in cropts['localrepos']: binddirs.append(lrepo) bindlist = map(expath, filter(None, binddirs)) bindlist += map(os.path.dirname, map(expath, filter(None, bindfiles))) bindlist = sorted(set(bindlist)) bindmounts = ';'.join(bindlist) return bindmounts def get_mic_binpath(): fp = None try: import pkg_resources # depends on 'setuptools' except ImportError: pass else: dist = pkg_resources.get_distribution('mic') # the real script is under EGG_INFO/scripts if dist.has_metadata('scripts/mic'): fp = os.path.join(dist.egg_info, "scripts/mic") if fp: return fp # not found script if 'flat' egg installed try: return find_binary_path('mic') except errors.CreatorError: raise errors.BootstrapError("Can't find mic binary in host OS") def get_mic_modpath(): try: import mic except ImportError: raise errors.BootstrapError("Can't find mic module in host OS") path = os.path.abspath(mic.__file__) return os.path.dirname(path) def get_mic_libpath(): # TBD: so far mic lib path is hard coded return "/usr/lib/mic" # the hard code path is prepared for bootstrap def sync_mic(bootstrap, binpth = '/usr/bin/mic', libpth='/usr/lib', pylib = '/usr/lib/python2.7/site-packages', conf = '/etc/mic/mic.conf'): _path = lambda p: os.path.join(bootstrap, p.lstrip('/')) micpaths = { 'binpth': get_mic_binpath(), 'libpth': get_mic_libpath(), 'pylib': get_mic_modpath(), 'conf': '/etc/mic/mic.conf', } if not os.path.exists(_path(pylib)): pyptn = '/usr/lib/python?.?/site-packages' pylibs = glob.glob(_path(pyptn)) if pylibs: pylib = pylibs[0].replace(bootstrap, '') else: raise errors.BootstrapError("Can't find python site dir in: %s" % bootstrap) for key, value in micpaths.items(): try: safecopy(value, _path(eval(key)), False, ["*.pyc", "*.pyo"]) except (OSError, IOError), err: raise errors.BootstrapError(err) # auto select backend conf_str = file(_path(conf)).read() conf_str = re.sub("pkgmgr\s*=\s*.*", "pkgmgr=auto", conf_str) with open(_path(conf), 'w') as wf: wf.write(conf_str) # chmod +x /usr/bin/mic os.chmod(_path(binpth), 0777) # correct python interpreter mic_cont = file(_path(binpth)).read() mic_cont = "#!/usr/bin/python\n" + mic_cont with open(_path(binpth), 'w') as wf: wf.write(mic_cont) def safecopy(src, dst, symlinks=False, ignore_ptns=()): if os.path.isdir(src): if os.path.isdir(dst): dst = os.path.join(dst, os.path.basename(src)) if os.path.exists(dst): shutil.rmtree(dst, ignore_errors=True) src = src.rstrip('/') # check common prefix to ignore copying itself if dst.startswith(src + '/'): ignore_ptns = list(ignore_ptns) + [ os.path.basename(src) ] ignores = shutil.ignore_patterns(*ignore_ptns) try: shutil.copytree(src, dst, symlinks, ignores) except (OSError, IOError): shutil.rmtree(dst, ignore_errors=True) raise else: if not os.path.isdir(dst): makedirs(os.path.dirname(dst)) shutil.copy2(src, dst) ########NEW FILE######## __FILENAME__ = BmapCreate """ This module implements the block map (bmap) creation functionality and provides the corresponding API in form of the 'BmapCreate' class. The idea is that while images files may generally be very large (e.g., 4GiB), they may nevertheless contain only little real data, e.g., 512MiB. This data are files, directories, file-system meta-data, partition table, etc. When copying the image to the target device, you do not have to copy all the 4GiB of data, you can copy only 512MiB of it, which is 4 times less, so copying should presumably be 4 times faster. The block map file is an XML file which contains a list of blocks which have to be copied to the target device. The other blocks are not used and there is no need to copy them. The XML file also contains some additional information like block size, image size, count of mapped blocks, etc. There are also many commentaries, so it is human-readable. The image has to be a sparse file. Generally, this means that when you generate this image file, you should start with a huge sparse file which contains a single hole spanning the entire file. Then you should partition it, write all the data (probably by means of loop-back mounting the image or parts of it), etc. The end result should be a sparse file where mapped areas represent useful parts of the image and holes represent useless parts of the image, which do not have to be copied when copying the image to the target device. This module uses the FIBMAP ioctl to detect holes. """ # Disable the following pylint recommendations: # * Too many instance attributes - R0902 # * Too few public methods - R0903 # pylint: disable=R0902,R0903 import hashlib from mic.utils.misc import human_size from mic.utils import Fiemap # The bmap format version we generate SUPPORTED_BMAP_VERSION = "1.3" _BMAP_START_TEMPLATE = \ """<?xml version="1.0" ?> <!-- This file contains the block map for an image file, which is basically a list of useful (mapped) block numbers in the image file. In other words, it lists only those blocks which contain data (boot sector, partition table, file-system metadata, files, directories, extents, etc). These blocks have to be copied to the target device. The other blocks do not contain any useful data and do not have to be copied to the target device. The block map an optimization which allows to copy or flash the image to the image quicker than copying of flashing the entire image. This is because with bmap less data is copied: <MappedBlocksCount> blocks instead of <BlocksCount> blocks. Besides the machine-readable data, this file contains useful commentaries which contain human-readable information like image size, percentage of mapped data, etc. The 'version' attribute is the block map file format version in the 'major.minor' format. The version major number is increased whenever an incompatible block map format change is made. The minor number changes in case of minor backward-compatible changes. --> <bmap version="%s"> <!-- Image size in bytes: %s --> <ImageSize> %u </ImageSize> <!-- Size of a block in bytes --> <BlockSize> %u </BlockSize> <!-- Count of blocks in the image file --> <BlocksCount> %u </BlocksCount> """ class Error(Exception): """ A class for exceptions generated by this module. We currently support only one type of exceptions, and we basically throw human-readable problem description in case of errors. """ pass class BmapCreate: """ This class implements the bmap creation functionality. To generate a bmap for an image (which is supposedly a sparse file), you should first create an instance of 'BmapCreate' and provide: * full path or a file-like object of the image to create bmap for * full path or a file object to use for writing the results to Then you should invoke the 'generate()' method of this class. It will use the FIEMAP ioctl to generate the bmap. """ def _open_image_file(self): """ Open the image file. """ try: self._f_image = open(self._image_path, 'rb') except IOError as err: raise Error("cannot open image file '%s': %s" \ % (self._image_path, err)) self._f_image_needs_close = True def _open_bmap_file(self): """ Open the bmap file. """ try: self._f_bmap = open(self._bmap_path, 'w+') except IOError as err: raise Error("cannot open bmap file '%s': %s" \ % (self._bmap_path, err)) self._f_bmap_needs_close = True def __init__(self, image, bmap): """ Initialize a class instance: * image - full path or a file-like object of the image to create bmap for * bmap - full path or a file object to use for writing the resulting bmap to """ self.image_size = None self.image_size_human = None self.block_size = None self.blocks_cnt = None self.mapped_cnt = None self.mapped_size = None self.mapped_size_human = None self.mapped_percent = None self._mapped_count_pos1 = None self._mapped_count_pos2 = None self._sha1_pos = None self._f_image_needs_close = False self._f_bmap_needs_close = False if hasattr(image, "read"): self._f_image = image self._image_path = image.name else: self._image_path = image self._open_image_file() if hasattr(bmap, "read"): self._f_bmap = bmap self._bmap_path = bmap.name else: self._bmap_path = bmap self._open_bmap_file() self.fiemap = Fiemap.Fiemap(self._f_image) self.image_size = self.fiemap.image_size self.image_size_human = human_size(self.image_size) if self.image_size == 0: raise Error("cannot generate bmap for zero-sized image file '%s'" \ % self._image_path) self.block_size = self.fiemap.block_size self.blocks_cnt = self.fiemap.blocks_cnt def _bmap_file_start(self): """ A helper function which generates the starting contents of the block map file: the header comment, image size, block size, etc. """ # We do not know the amount of mapped blocks at the moment, so just put # whitespaces instead of real numbers. Assume the longest possible # numbers. mapped_count = ' ' * len(str(self.image_size)) mapped_size_human = ' ' * len(self.image_size_human) xml = _BMAP_START_TEMPLATE \ % (SUPPORTED_BMAP_VERSION, self.image_size_human, self.image_size, self.block_size, self.blocks_cnt) xml += " <!-- Count of mapped blocks: " self._f_bmap.write(xml) self._mapped_count_pos1 = self._f_bmap.tell() # Just put white-spaces instead of real information about mapped blocks xml = "%s or %.1f -->\n" % (mapped_size_human, 100.0) xml += " <MappedBlocksCount> " self._f_bmap.write(xml) self._mapped_count_pos2 = self._f_bmap.tell() xml = "%s </MappedBlocksCount>\n\n" % mapped_count # pylint: disable=C0301 xml += " <!-- The checksum of this bmap file. When it is calculated, the value of\n" xml += " the SHA1 checksum has be zeoro (40 ASCII \"0\" symbols). -->\n" xml += " <BmapFileSHA1> " self._f_bmap.write(xml) self._sha1_pos = self._f_bmap.tell() xml = "0" * 40 + " </BmapFileSHA1>\n\n" xml += " <!-- The block map which consists of elements which may either be a\n" xml += " range of blocks or a single block. The 'sha1' attribute (if present)\n" xml += " is the SHA1 checksum of this blocks range. -->\n" xml += " <BlockMap>\n" # pylint: enable=C0301 self._f_bmap.write(xml) def _bmap_file_end(self): """ A helper function which generates the final parts of the block map file: the ending tags and the information about the amount of mapped blocks. """ xml = " </BlockMap>\n" xml += "</bmap>\n" self._f_bmap.write(xml) self._f_bmap.seek(self._mapped_count_pos1) self._f_bmap.write("%s or %.1f%%" % \ (self.mapped_size_human, self.mapped_percent)) self._f_bmap.seek(self._mapped_count_pos2) self._f_bmap.write("%u" % self.mapped_cnt) self._f_bmap.seek(0) sha1 = hashlib.sha1(self._f_bmap.read()).hexdigest() self._f_bmap.seek(self._sha1_pos) self._f_bmap.write("%s" % sha1) def _calculate_sha1(self, first, last): """ A helper function which calculates SHA1 checksum for the range of blocks of the image file: from block 'first' to block 'last'. """ start = first * self.block_size end = (last + 1) * self.block_size self._f_image.seek(start) hash_obj = hashlib.new("sha1") chunk_size = 1024*1024 to_read = end - start read = 0 while read < to_read: if read + chunk_size > to_read: chunk_size = to_read - read chunk = self._f_image.read(chunk_size) hash_obj.update(chunk) read += chunk_size return hash_obj.hexdigest() def generate(self, include_checksums = True): """ Generate bmap for the image file. If 'include_checksums' is 'True', also generate SHA1 checksums for block ranges. """ # Save image file position in order to restore it at the end image_pos = self._f_image.tell() self._bmap_file_start() # Generate the block map and write it to the XML block map # file as we go. self.mapped_cnt = 0 for first, last in self.fiemap.get_mapped_ranges(0, self.blocks_cnt): self.mapped_cnt += last - first + 1 if include_checksums: sha1 = self._calculate_sha1(first, last) sha1 = " sha1=\"%s\"" % sha1 else: sha1 = "" if first != last: self._f_bmap.write(" <Range%s> %s-%s </Range>\n" \ % (sha1, first, last)) else: self._f_bmap.write(" <Range%s> %s </Range>\n" \ % (sha1, first)) self.mapped_size = self.mapped_cnt * self.block_size self.mapped_size_human = human_size(self.mapped_size) self.mapped_percent = (self.mapped_cnt * 100.0) / self.blocks_cnt self._bmap_file_end() try: self._f_bmap.flush() except IOError as err: raise Error("cannot flush the bmap file '%s': %s" \ % (self._bmap_path, err)) self._f_image.seek(image_pos) def __del__(self): """ The class destructor which closes the opened files. """ if self._f_image_needs_close: self._f_image.close() if self._f_bmap_needs_close: self._f_bmap.close() ########NEW FILE######## __FILENAME__ = cmdln #!/usr/bin/env python # Copyright (c) 2002-2007 ActiveState Software Inc. # License: MIT (see LICENSE.txt for license details) # Author: Trent Mick # Home: http://trentm.com/projects/cmdln/ """An improvement on Python's standard cmd.py module. As with cmd.py, this module provides "a simple framework for writing line-oriented command intepreters." This module provides a 'RawCmdln' class that fixes some design flaws in cmd.Cmd, making it more scalable and nicer to use for good 'cvs'- or 'svn'-style command line interfaces or simple shells. And it provides a 'Cmdln' class that add optparse-based option processing. Basically you use it like this: import cmdln class MySVN(cmdln.Cmdln): name = "svn" @cmdln.alias('stat', 'st') @cmdln.option('-v', '--verbose', action='store_true' help='print verbose information') def do_status(self, subcmd, opts, *paths): print "handle 'svn status' command" #... if __name__ == "__main__": shell = MySVN() retval = shell.main() sys.exit(retval) See the README.txt or <http://trentm.com/projects/cmdln/> for more details. """ __version_info__ = (1, 1, 2) __version__ = '.'.join(map(str, __version_info__)) import os import sys import re import cmd import optparse from pprint import pprint import sys #---- globals LOOP_ALWAYS, LOOP_NEVER, LOOP_IF_EMPTY = range(3) # An unspecified optional argument when None is a meaningful value. _NOT_SPECIFIED = ("Not", "Specified") # Pattern to match a TypeError message from a call that # failed because of incorrect number of arguments (see # Python/getargs.c). _INCORRECT_NUM_ARGS_RE = re.compile( r"(takes [\w ]+ )(\d+)( arguments? \()(\d+)( given\))") #---- exceptions class CmdlnError(Exception): """A cmdln.py usage error.""" def __init__(self, msg): self.msg = msg def __str__(self): return self.msg class CmdlnUserError(Exception): """An error by a user of a cmdln-based tool/shell.""" pass #---- public methods and classes def alias(*aliases): """Decorator to add aliases for Cmdln.do_* command handlers. Example: class MyShell(cmdln.Cmdln): @cmdln.alias("!", "sh") def do_shell(self, argv): #...implement 'shell' command """ def decorate(f): if not hasattr(f, "aliases"): f.aliases = [] f.aliases += aliases return f return decorate class RawCmdln(cmd.Cmd): """An improved (on cmd.Cmd) framework for building multi-subcommand scripts (think "svn" & "cvs") and simple shells (think "pdb" and "gdb"). A simple example: import cmdln class MySVN(cmdln.RawCmdln): name = "svn" @cmdln.aliases('stat', 'st') def do_status(self, argv): print "handle 'svn status' command" if __name__ == "__main__": shell = MySVN() retval = shell.main() sys.exit(retval) See <http://trentm.com/projects/cmdln> for more information. """ name = None # if unset, defaults basename(sys.argv[0]) prompt = None # if unset, defaults to self.name+"> " version = None # if set, default top-level options include --version # Default messages for some 'help' command error cases. # They are interpolated with one arg: the command. nohelp = "no help on '%s'" unknowncmd = "unknown command: '%s'" helpindent = '' # string with which to indent help output def __init__(self, completekey='tab', stdin=None, stdout=None, stderr=None): """Cmdln(completekey='tab', stdin=None, stdout=None, stderr=None) The optional argument 'completekey' is the readline name of a completion key; it defaults to the Tab key. If completekey is not None and the readline module is available, command completion is done automatically. The optional arguments 'stdin', 'stdout' and 'stderr' specify alternate input, output and error output file objects; if not specified, sys.* are used. If 'stdout' but not 'stderr' is specified, stdout is used for error output. This is to provide least surprise for users used to only the 'stdin' and 'stdout' options with cmd.Cmd. """ import sys if self.name is None: self.name = os.path.basename(sys.argv[0]) if self.prompt is None: self.prompt = self.name+"> " self._name_str = self._str(self.name) self._prompt_str = self._str(self.prompt) if stdin is not None: self.stdin = stdin else: self.stdin = sys.stdin if stdout is not None: self.stdout = stdout else: self.stdout = sys.stdout if stderr is not None: self.stderr = stderr elif stdout is not None: self.stderr = stdout else: self.stderr = sys.stderr self.cmdqueue = [] self.completekey = completekey self.cmdlooping = False def get_optparser(self): """Hook for subclasses to set the option parser for the top-level command/shell. This option parser is used retrieved and used by `.main()' to handle top-level options. The default implements a single '-h|--help' option. Sub-classes can return None to have no options at the top-level. Typically an instance of CmdlnOptionParser should be returned. """ version = (self.version is not None and "%s %s" % (self._name_str, self.version) or None) return CmdlnOptionParser(self, version=version) def postoptparse(self): """Hook method executed just after `.main()' parses top-level options. When called `self.options' holds the results of the option parse. """ pass def main(self, argv=None, loop=LOOP_NEVER): """A possible mainline handler for a script, like so: import cmdln class MyCmd(cmdln.Cmdln): name = "mycmd" ... if __name__ == "__main__": MyCmd().main() By default this will use sys.argv to issue a single command to 'MyCmd', then exit. The 'loop' argument can be use to control interactive shell behaviour. Arguments: "argv" (optional, default sys.argv) is the command to run. It must be a sequence, where the first element is the command name and subsequent elements the args for that command. "loop" (optional, default LOOP_NEVER) is a constant indicating if a command loop should be started (i.e. an interactive shell). Valid values (constants on this module): LOOP_ALWAYS start loop and run "argv", if any LOOP_NEVER run "argv" (or .emptyline()) and exit LOOP_IF_EMPTY run "argv", if given, and exit; otherwise, start loop """ if argv is None: import sys argv = sys.argv else: argv = argv[:] # don't modify caller's list self.optparser = self.get_optparser() if self.optparser: # i.e. optparser=None means don't process for opts try: self.options, args = self.optparser.parse_args(argv[1:]) except CmdlnUserError, ex: msg = "%s: %s\nTry '%s help' for info.\n"\ % (self.name, ex, self.name) self.stderr.write(self._str(msg)) self.stderr.flush() return 1 except StopOptionProcessing, ex: return 0 else: self.options, args = None, argv[1:] self.postoptparse() if loop == LOOP_ALWAYS: if args: self.cmdqueue.append(args) return self.cmdloop() elif loop == LOOP_NEVER: if args: return self.cmd(args) else: return self.emptyline() elif loop == LOOP_IF_EMPTY: if args: return self.cmd(args) else: return self.cmdloop() def cmd(self, argv): """Run one command and exit. "argv" is the arglist for the command to run. argv[0] is the command to run. If argv is an empty list then the 'emptyline' handler is run. Returns the return value from the command handler. """ assert isinstance(argv, (list, tuple)), \ "'argv' is not a sequence: %r" % argv retval = None try: argv = self.precmd(argv) retval = self.onecmd(argv) self.postcmd(argv) except: if not self.cmdexc(argv): raise retval = 1 return retval def _str(self, s): """Safely convert the given str/unicode to a string for printing.""" try: return str(s) except UnicodeError: #XXX What is the proper encoding to use here? 'utf-8' seems # to work better than "getdefaultencoding" (usually # 'ascii'), on OS X at least. #import sys #return s.encode(sys.getdefaultencoding(), "replace") return s.encode("utf-8", "replace") def cmdloop(self, intro=None): """Repeatedly issue a prompt, accept input, parse into an argv, and dispatch (via .precmd(), .onecmd() and .postcmd()), passing them the argv. In other words, start a shell. "intro" (optional) is a introductory message to print when starting the command loop. This overrides the class "intro" attribute, if any. """ self.cmdlooping = True self.preloop() if self.use_rawinput and self.completekey: try: import readline self.old_completer = readline.get_completer() readline.set_completer(self.complete) readline.parse_and_bind(self.completekey+": complete") except ImportError: pass try: if intro is None: intro = self.intro if intro: intro_str = self._str(intro) self.stdout.write(intro_str+'\n') self.stop = False retval = None while not self.stop: if self.cmdqueue: argv = self.cmdqueue.pop(0) assert isinstance(argv, (list, tuple)), \ "item on 'cmdqueue' is not a sequence: %r" % argv else: if self.use_rawinput: try: line = raw_input(self._prompt_str) except EOFError: line = 'EOF' else: self.stdout.write(self._prompt_str) self.stdout.flush() line = self.stdin.readline() if not len(line): line = 'EOF' else: line = line[:-1] # chop '\n' argv = line2argv(line) try: argv = self.precmd(argv) retval = self.onecmd(argv) self.postcmd(argv) except: if not self.cmdexc(argv): raise retval = 1 self.lastretval = retval self.postloop() finally: if self.use_rawinput and self.completekey: try: import readline readline.set_completer(self.old_completer) except ImportError: pass self.cmdlooping = False return retval def precmd(self, argv): """Hook method executed just before the command argv is interpreted, but after the input prompt is generated and issued. "argv" is the cmd to run. Returns an argv to run (i.e. this method can modify the command to run). """ return argv def postcmd(self, argv): """Hook method executed just after a command dispatch is finished. "argv" is the command that was run. """ pass def cmdexc(self, argv): """Called if an exception is raised in any of precmd(), onecmd(), or postcmd(). If True is returned, the exception is deemed to have been dealt with. Otherwise, the exception is re-raised. The default implementation handles CmdlnUserError's, which typically correspond to user error in calling commands (as opposed to programmer error in the design of the script using cmdln.py). """ import sys type, exc, traceback = sys.exc_info() if isinstance(exc, CmdlnUserError): msg = "%s %s: %s\nTry '%s help %s' for info.\n"\ % (self.name, argv[0], exc, self.name, argv[0]) self.stderr.write(self._str(msg)) self.stderr.flush() return True def onecmd(self, argv): if not argv: return self.emptyline() self.lastcmd = argv cmdname = self._get_canonical_cmd_name(argv[0]) if cmdname: handler = self._get_cmd_handler(cmdname) if handler: return self._dispatch_cmd(handler, argv) return self.default(argv) def _dispatch_cmd(self, handler, argv): return handler(argv) def default(self, argv): """Hook called to handle a command for which there is no handler. "argv" is the command and arguments to run. The default implementation writes and error message to stderr and returns an error exit status. Returns a numeric command exit status. """ errmsg = self._str(self.unknowncmd % (argv[0],)) if self.cmdlooping: self.stderr.write(errmsg+"\n") else: self.stderr.write("%s: %s\nTry '%s help' for info.\n" % (self._name_str, errmsg, self._name_str)) self.stderr.flush() return 1 def parseline(self, line): # This is used by Cmd.complete (readline completer function) to # massage the current line buffer before completion processing. # We override to drop special '!' handling. line = line.strip() if not line: return None, None, line elif line[0] == '?': line = 'help ' + line[1:] i, n = 0, len(line) while i < n and line[i] in self.identchars: i = i+1 cmd, arg = line[:i], line[i:].strip() return cmd, arg, line def helpdefault(self, cmd, known): """Hook called to handle help on a command for which there is no help handler. "cmd" is the command name on which help was requested. "known" is a boolean indicating if this command is known (i.e. if there is a handler for it). Returns a return code. """ if known: msg = self._str(self.nohelp % (cmd,)) if self.cmdlooping: self.stderr.write(msg + '\n') else: self.stderr.write("%s: %s\n" % (self.name, msg)) else: msg = self.unknowncmd % (cmd,) if self.cmdlooping: self.stderr.write(msg + '\n') else: self.stderr.write("%s: %s\n" "Try '%s help' for info.\n" % (self.name, msg, self.name)) self.stderr.flush() return 1 def do_help(self, argv): """${cmd_name}: give detailed help on a specific sub-command Usage: ${name} help [COMMAND] """ if len(argv) > 1: # asking for help on a particular command doc = None cmdname = self._get_canonical_cmd_name(argv[1]) or argv[1] if not cmdname: return self.helpdefault(argv[1], False) else: helpfunc = getattr(self, "help_"+cmdname, None) if helpfunc: doc = helpfunc() else: handler = self._get_cmd_handler(cmdname) if handler: doc = handler.__doc__ if doc is None: return self.helpdefault(argv[1], handler != None) else: # bare "help" command doc = self.__class__.__doc__ # try class docstring if doc is None: # Try to provide some reasonable useful default help. if self.cmdlooping: prefix = "" else: prefix = self.name+' ' doc = """Usage: %sCOMMAND [ARGS...] %shelp [COMMAND] ${option_list} ${command_list} ${help_list} """ % (prefix, prefix) cmdname = None if doc: # *do* have help content, massage and print that doc = self._help_reindent(doc) doc = self._help_preprocess(doc, cmdname) doc = doc.rstrip() + '\n' # trim down trailing space self.stdout.write(self._str(doc)) self.stdout.flush() do_help.aliases = ["?"] def _help_reindent(self, help, indent=None): """Hook to re-indent help strings before writing to stdout. "help" is the help content to re-indent "indent" is a string with which to indent each line of the help content after normalizing. If unspecified or None then the default is use: the 'self.helpindent' class attribute. By default this is the empty string, i.e. no indentation. By default, all common leading whitespace is removed and then the lot is indented by 'self.helpindent'. When calculating the common leading whitespace the first line is ignored -- hence help content for Conan can be written as follows and have the expected indentation: def do_crush(self, ...): '''${cmd_name}: crush your enemies, see them driven before you... c.f. Conan the Barbarian''' """ if indent is None: indent = self.helpindent lines = help.splitlines(0) _dedentlines(lines, skip_first_line=True) lines = [(indent+line).rstrip() for line in lines] return '\n'.join(lines) def _help_preprocess(self, help, cmdname): """Hook to preprocess a help string before writing to stdout. "help" is the help string to process. "cmdname" is the canonical sub-command name for which help is being given, or None if the help is not specific to a command. By default the following template variables are interpolated in help content. (Note: these are similar to Python 2.4's string.Template interpolation but not quite.) ${name} The tool's/shell's name, i.e. 'self.name'. ${option_list} A formatted table of options for this shell/tool. ${command_list} A formatted table of available sub-commands. ${help_list} A formatted table of additional help topics (i.e. 'help_*' methods with no matching 'do_*' method). ${cmd_name} The name (and aliases) for this sub-command formatted as: "NAME (ALIAS1, ALIAS2, ...)". ${cmd_usage} A formatted usage block inferred from the command function signature. ${cmd_option_list} A formatted table of options for this sub-command. (This is only available for commands using the optparse integration, i.e. using @cmdln.option decorators or manually setting the 'optparser' attribute on the 'do_*' method.) Returns the processed help. """ preprocessors = { "${name}": self._help_preprocess_name, "${option_list}": self._help_preprocess_option_list, "${command_list}": self._help_preprocess_command_list, "${help_list}": self._help_preprocess_help_list, "${cmd_name}": self._help_preprocess_cmd_name, "${cmd_usage}": self._help_preprocess_cmd_usage, "${cmd_option_list}": self._help_preprocess_cmd_option_list, } for marker, preprocessor in preprocessors.items(): if marker in help: help = preprocessor(help, cmdname) return help def _help_preprocess_name(self, help, cmdname=None): return help.replace("${name}", self.name) def _help_preprocess_option_list(self, help, cmdname=None): marker = "${option_list}" indent, indent_width = _get_indent(marker, help) suffix = _get_trailing_whitespace(marker, help) if self.optparser: # Setup formatting options and format. # - Indentation of 4 is better than optparse default of 2. # C.f. Damian Conway's discussion of this in Perl Best # Practices. self.optparser.formatter.indent_increment = 4 self.optparser.formatter.current_indent = indent_width block = self.optparser.format_option_help() + '\n' else: block = "" help = help.replace(indent+marker+suffix, block, 1) return help def _help_preprocess_command_list(self, help, cmdname=None): marker = "${command_list}" indent, indent_width = _get_indent(marker, help) suffix = _get_trailing_whitespace(marker, help) # Find any aliases for commands. token2canonical = self._get_canonical_map() aliases = {} for token, cmdname in token2canonical.items(): if token == cmdname: continue aliases.setdefault(cmdname, []).append(token) # Get the list of (non-hidden) commands and their # documentation, if any. cmdnames = {} # use a dict to strip duplicates for attr in self.get_names(): if attr.startswith("do_"): cmdnames[attr[3:]] = True cmdnames = cmdnames.keys() cmdnames.sort() linedata = [] for cmdname in cmdnames: if aliases.get(cmdname): a = aliases[cmdname] a.sort() cmdstr = "%s (%s)" % (cmdname, ", ".join(a)) else: cmdstr = cmdname doc = None try: helpfunc = getattr(self, 'help_'+cmdname) except AttributeError: handler = self._get_cmd_handler(cmdname) if handler: doc = handler.__doc__ else: doc = helpfunc() # Strip "${cmd_name}: " from the start of a command's doc. Best # practice dictates that command help strings begin with this, but # it isn't at all wanted for the command list. to_strip = "${cmd_name}:" if doc and doc.startswith(to_strip): #log.debug("stripping %r from start of %s's help string", # to_strip, cmdname) doc = doc[len(to_strip):].lstrip() linedata.append( (cmdstr, doc) ) if linedata: subindent = indent + ' '*4 lines = _format_linedata(linedata, subindent, indent_width+4) block = indent + "Commands:\n" \ + '\n'.join(lines) + "\n\n" help = help.replace(indent+marker+suffix, block, 1) return help def _gen_names_and_attrs(self): # Inheritance says we have to look in class and # base classes; order is not important. names = [] classes = [self.__class__] while classes: aclass = classes.pop(0) if aclass.__bases__: classes = classes + list(aclass.__bases__) for name in dir(aclass): yield (name, getattr(aclass, name)) def _help_preprocess_help_list(self, help, cmdname=None): marker = "${help_list}" indent, indent_width = _get_indent(marker, help) suffix = _get_trailing_whitespace(marker, help) # Determine the additional help topics, if any. helpnames = {} token2cmdname = self._get_canonical_map() for attrname, attr in self._gen_names_and_attrs(): if not attrname.startswith("help_"): continue helpname = attrname[5:] if helpname not in token2cmdname: helpnames[helpname] = attr if helpnames: linedata = [(n, a.__doc__ or "") for n, a in helpnames.items()] linedata.sort() subindent = indent + ' '*4 lines = _format_linedata(linedata, subindent, indent_width+4) block = (indent + "Additional help topics (run `%s help TOPIC'):\n" % self.name + '\n'.join(lines) + "\n\n") else: block = '' help = help.replace(indent+marker+suffix, block, 1) return help def _help_preprocess_cmd_name(self, help, cmdname=None): marker = "${cmd_name}" handler = self._get_cmd_handler(cmdname) if not handler: raise CmdlnError("cannot preprocess '%s' into help string: " "could not find command handler for %r" % (marker, cmdname)) s = cmdname if hasattr(handler, "aliases"): s += " (%s)" % (", ".join(handler.aliases)) help = help.replace(marker, s) return help #TODO: this only makes sense as part of the Cmdln class. # Add hooks to add help preprocessing template vars and put # this one on that class. def _help_preprocess_cmd_usage(self, help, cmdname=None): marker = "${cmd_usage}" handler = self._get_cmd_handler(cmdname) if not handler: raise CmdlnError("cannot preprocess '%s' into help string: " "could not find command handler for %r" % (marker, cmdname)) indent, indent_width = _get_indent(marker, help) suffix = _get_trailing_whitespace(marker, help) # Extract the introspection bits we need. func = handler.im_func if func.func_defaults: func_defaults = list(func.func_defaults) else: func_defaults = [] co_argcount = func.func_code.co_argcount co_varnames = func.func_code.co_varnames co_flags = func.func_code.co_flags CO_FLAGS_ARGS = 4 CO_FLAGS_KWARGS = 8 # Adjust argcount for possible *args and **kwargs arguments. argcount = co_argcount if co_flags & CO_FLAGS_ARGS: argcount += 1 if co_flags & CO_FLAGS_KWARGS: argcount += 1 # Determine the usage string. usage = "%s %s" % (self.name, cmdname) if argcount <= 2: # handler ::= do_FOO(self, argv) usage += " [ARGS...]" elif argcount >= 3: # handler ::= do_FOO(self, subcmd, opts, ...) argnames = list(co_varnames[3:argcount]) tail = "" if co_flags & CO_FLAGS_KWARGS: name = argnames.pop(-1) import warnings # There is no generally accepted mechanism for passing # keyword arguments from the command line. Could # *perhaps* consider: arg=value arg2=value2 ... warnings.warn("argument '**%s' on '%s.%s' command " "handler will never get values" % (name, self.__class__.__name__, func.func_name)) if co_flags & CO_FLAGS_ARGS: name = argnames.pop(-1) tail = "[%s...]" % name.upper() while func_defaults: func_defaults.pop(-1) name = argnames.pop(-1) tail = "[%s%s%s]" % (name.upper(), (tail and ' ' or ''), tail) while argnames: name = argnames.pop(-1) tail = "%s %s" % (name.upper(), tail) usage += ' ' + tail block_lines = [ self.helpindent + "Usage:", self.helpindent + ' '*4 + usage ] block = '\n'.join(block_lines) + '\n\n' help = help.replace(indent+marker+suffix, block, 1) return help #TODO: this only makes sense as part of the Cmdln class. # Add hooks to add help preprocessing template vars and put # this one on that class. def _help_preprocess_cmd_option_list(self, help, cmdname=None): marker = "${cmd_option_list}" handler = self._get_cmd_handler(cmdname) if not handler: raise CmdlnError("cannot preprocess '%s' into help string: " "could not find command handler for %r" % (marker, cmdname)) indent, indent_width = _get_indent(marker, help) suffix = _get_trailing_whitespace(marker, help) if hasattr(handler, "optparser"): # Setup formatting options and format. # - Indentation of 4 is better than optparse default of 2. # C.f. Damian Conway's discussion of this in Perl Best # Practices. handler.optparser.formatter.indent_increment = 4 handler.optparser.formatter.current_indent = indent_width block = handler.optparser.format_option_help() + '\n' else: block = "" help = help.replace(indent+marker+suffix, block, 1) return help def _get_canonical_cmd_name(self, token): map = self._get_canonical_map() return map.get(token, None) def _get_canonical_map(self): """Return a mapping of available command names and aliases to their canonical command name. """ cacheattr = "_token2canonical" if not hasattr(self, cacheattr): # Get the list of commands and their aliases, if any. token2canonical = {} cmd2funcname = {} # use a dict to strip duplicates for attr in self.get_names(): if attr.startswith("do_"): cmdname = attr[3:] elif attr.startswith("_do_"): cmdname = attr[4:] else: continue cmd2funcname[cmdname] = attr token2canonical[cmdname] = cmdname for cmdname, funcname in cmd2funcname.items(): # add aliases func = getattr(self, funcname) aliases = getattr(func, "aliases", []) for alias in aliases: if alias in cmd2funcname: import warnings warnings.warn("'%s' alias for '%s' command conflicts " "with '%s' handler" % (alias, cmdname, cmd2funcname[alias])) continue token2canonical[alias] = cmdname setattr(self, cacheattr, token2canonical) return getattr(self, cacheattr) def _get_cmd_handler(self, cmdname): handler = None try: handler = getattr(self, 'do_' + cmdname) except AttributeError: try: # Private command handlers begin with "_do_". handler = getattr(self, '_do_' + cmdname) except AttributeError: pass return handler def _do_EOF(self, argv): # Default EOF handler # Note: an actual EOF is redirected to this command. #TODO: separate name for this. Currently it is available from # command-line. Is that okay? self.stdout.write('\n') self.stdout.flush() self.stop = True def emptyline(self): # Different from cmd.Cmd: don't repeat the last command for an # emptyline. if self.cmdlooping: pass else: return self.do_help(["help"]) #---- optparse.py extension to fix (IMO) some deficiencies # # See the class _OptionParserEx docstring for details. # class StopOptionProcessing(Exception): """Indicate that option *and argument* processing should stop cleanly. This is not an error condition. It is similar in spirit to StopIteration. This is raised by _OptionParserEx's default "help" and "version" option actions and can be raised by custom option callbacks too. Hence the typical CmdlnOptionParser (a subclass of _OptionParserEx) usage is: parser = CmdlnOptionParser(mycmd) parser.add_option("-f", "--force", dest="force") ... try: opts, args = parser.parse_args() except StopOptionProcessing: # normal termination, "--help" was probably given sys.exit(0) """ class _OptionParserEx(optparse.OptionParser): """An optparse.OptionParser that uses exceptions instead of sys.exit. This class is an extension of optparse.OptionParser that differs as follows: - Correct (IMO) the default OptionParser error handling to never sys.exit(). Instead OptParseError exceptions are passed through. - Add the StopOptionProcessing exception (a la StopIteration) to indicate normal termination of option processing. See StopOptionProcessing's docstring for details. I'd also like to see the following in the core optparse.py, perhaps as a RawOptionParser which would serve as a base class for the more generally used OptionParser (that works as current): - Remove the implicit addition of the -h|--help and --version options. They can get in the way (e.g. if want '-?' and '-V' for these as well) and it is not hard to do: optparser.add_option("-h", "--help", action="help") optparser.add_option("--version", action="version") These are good practices, just not valid defaults if they can get in the way. """ def error(self, msg): raise optparse.OptParseError(msg) def exit(self, status=0, msg=None): if status == 0: raise StopOptionProcessing(msg) else: #TODO: don't lose status info here raise optparse.OptParseError(msg) #---- optparse.py-based option processing support class CmdlnOptionParser(_OptionParserEx): """An optparse.OptionParser class more appropriate for top-level Cmdln options. For parsing of sub-command options, see SubCmdOptionParser. Changes: - disable_interspersed_args() by default, because a Cmdln instance has sub-commands which may themselves have options. - Redirect print_help() to the Cmdln.do_help() which is better equiped to handle the "help" action. - error() will raise a CmdlnUserError: OptionParse.error() is meant to be called for user errors. Raising a well-known error here can make error handling clearer. - Also see the changes in _OptionParserEx. """ def __init__(self, cmdln, **kwargs): self.cmdln = cmdln kwargs["prog"] = self.cmdln.name _OptionParserEx.__init__(self, **kwargs) self.disable_interspersed_args() def print_help(self, file=None): self.cmdln.onecmd(["help"]) def error(self, msg): raise CmdlnUserError(msg) class SubCmdOptionParser(_OptionParserEx): def set_cmdln_info(self, cmdln, subcmd): """Called by Cmdln to pass relevant info about itself needed for print_help(). """ self.cmdln = cmdln self.subcmd = subcmd def print_help(self, file=None): self.cmdln.onecmd(["help", self.subcmd]) def error(self, msg): raise CmdlnUserError(msg) def option(*args, **kwargs): """Decorator to add an option to the optparser argument of a Cmdln subcommand. Example: class MyShell(cmdln.Cmdln): @cmdln.option("-f", "--force", help="force removal") def do_remove(self, subcmd, opts, *args): #... """ #XXX Is there a possible optimization for many options to not have a # large stack depth here? def decorate(f): if not hasattr(f, "optparser"): f.optparser = SubCmdOptionParser() f.optparser.add_option(*args, **kwargs) return f return decorate class Cmdln(RawCmdln): """An improved (on cmd.Cmd) framework for building multi-subcommand scripts (think "svn" & "cvs") and simple shells (think "pdb" and "gdb"). A simple example: import cmdln class MySVN(cmdln.Cmdln): name = "svn" @cmdln.aliases('stat', 'st') @cmdln.option('-v', '--verbose', action='store_true' help='print verbose information') def do_status(self, subcmd, opts, *paths): print "handle 'svn status' command" #... if __name__ == "__main__": shell = MySVN() retval = shell.main() sys.exit(retval) 'Cmdln' extends 'RawCmdln' by providing optparse option processing integration. See this class' _dispatch_cmd() docstring and <http://trentm.com/projects/cmdln> for more information. """ def _dispatch_cmd(self, handler, argv): """Introspect sub-command handler signature to determine how to dispatch the command. The raw handler provided by the base 'RawCmdln' class is still supported: def do_foo(self, argv): # 'argv' is the vector of command line args, argv[0] is # the command name itself (i.e. "foo" or an alias) pass In addition, if the handler has more than 2 arguments option processing is automatically done (using optparse): @cmdln.option('-v', '--verbose', action='store_true') def do_bar(self, subcmd, opts, *args): # subcmd = <"bar" or an alias> # opts = <an optparse.Values instance> if opts.verbose: print "lots of debugging output..." # args = <tuple of arguments> for arg in args: bar(arg) TODO: explain that "*args" can be other signatures as well. The `cmdln.option` decorator corresponds to an `add_option()` method call on an `optparse.OptionParser` instance. You can declare a specific number of arguments: @cmdln.option('-v', '--verbose', action='store_true') def do_bar2(self, subcmd, opts, bar_one, bar_two): #... and an appropriate error message will be raised/printed if the command is called with a different number of args. """ co_argcount = handler.im_func.func_code.co_argcount if co_argcount == 2: # handler ::= do_foo(self, argv) return handler(argv) elif co_argcount >= 3: # handler ::= do_foo(self, subcmd, opts, ...) try: optparser = handler.optparser except AttributeError: optparser = handler.im_func.optparser = SubCmdOptionParser() assert isinstance(optparser, SubCmdOptionParser) optparser.set_cmdln_info(self, argv[0]) try: opts, args = optparser.parse_args(argv[1:]) except StopOptionProcessing: #TODO: this doesn't really fly for a replacement of # optparse.py behaviour, does it? return 0 # Normal command termination try: return handler(argv[0], opts, *args) except TypeError, ex: # Some TypeError's are user errors: # do_foo() takes at least 4 arguments (3 given) # do_foo() takes at most 5 arguments (6 given) # do_foo() takes exactly 5 arguments (6 given) # Raise CmdlnUserError for these with a suitably # massaged error message. import sys tb = sys.exc_info()[2] # the traceback object if tb.tb_next is not None: # If the traceback is more than one level deep, then the # TypeError do *not* happen on the "handler(...)" call # above. In that we don't want to handle it specially # here: it would falsely mask deeper code errors. raise msg = ex.args[0] match = _INCORRECT_NUM_ARGS_RE.search(msg) if match: msg = list(match.groups()) msg[1] = int(msg[1]) - 3 if msg[1] == 1: msg[2] = msg[2].replace("arguments", "argument") msg[3] = int(msg[3]) - 3 msg = ''.join(map(str, msg)) raise CmdlnUserError(msg) else: raise else: raise CmdlnError("incorrect argcount for %s(): takes %d, must " "take 2 for 'argv' signature or 3+ for 'opts' " "signature" % (handler.__name__, co_argcount)) #---- internal support functions def _format_linedata(linedata, indent, indent_width): """Format specific linedata into a pleasant layout. "linedata" is a list of 2-tuples of the form: (<item-display-string>, <item-docstring>) "indent" is a string to use for one level of indentation "indent_width" is a number of columns by which the formatted data will be indented when printed. The <item-display-string> column is held to 15 columns. """ lines = [] WIDTH = 78 - indent_width SPACING = 2 NAME_WIDTH_LOWER_BOUND = 13 NAME_WIDTH_UPPER_BOUND = 16 NAME_WIDTH = max([len(s) for s,d in linedata]) if NAME_WIDTH < NAME_WIDTH_LOWER_BOUND: NAME_WIDTH = NAME_WIDTH_LOWER_BOUND else: NAME_WIDTH = NAME_WIDTH_UPPER_BOUND DOC_WIDTH = WIDTH - NAME_WIDTH - SPACING for namestr, doc in linedata: line = indent + namestr if len(namestr) <= NAME_WIDTH: line += ' ' * (NAME_WIDTH + SPACING - len(namestr)) else: lines.append(line) line = indent + ' ' * (NAME_WIDTH + SPACING) line += _summarize_doc(doc, DOC_WIDTH) lines.append(line.rstrip()) return lines def _summarize_doc(doc, length=60): r"""Parse out a short one line summary from the given doclines. "doc" is the doc string to summarize. "length" is the max length for the summary >>> _summarize_doc("this function does this") 'this function does this' >>> _summarize_doc("this function does this", 10) 'this fu...' >>> _summarize_doc("this function does this\nand that") 'this function does this and that' >>> _summarize_doc("this function does this\n\nand that") 'this function does this' """ import re if doc is None: return "" assert length > 3, "length <= 3 is absurdly short for a doc summary" doclines = doc.strip().splitlines(0) if not doclines: return "" summlines = [] for i, line in enumerate(doclines): stripped = line.strip() if not stripped: break summlines.append(stripped) if len(''.join(summlines)) >= length: break summary = ' '.join(summlines) if len(summary) > length: summary = summary[:length-3] + "..." return summary def line2argv(line): r"""Parse the given line into an argument vector. "line" is the line of input to parse. This may get niggly when dealing with quoting and escaping. The current state of this parsing may not be completely thorough/correct in this respect. >>> from cmdln import line2argv >>> line2argv("foo") ['foo'] >>> line2argv("foo bar") ['foo', 'bar'] >>> line2argv("foo bar ") ['foo', 'bar'] >>> line2argv(" foo bar") ['foo', 'bar'] Quote handling: >>> line2argv("'foo bar'") ['foo bar'] >>> line2argv('"foo bar"') ['foo bar'] >>> line2argv(r'"foo\"bar"') ['foo"bar'] >>> line2argv("'foo bar' spam") ['foo bar', 'spam'] >>> line2argv("'foo 'bar spam") ['foo bar', 'spam'] >>> line2argv('some\tsimple\ttests') ['some', 'simple', 'tests'] >>> line2argv('a "more complex" test') ['a', 'more complex', 'test'] >>> line2argv('a more="complex test of " quotes') ['a', 'more=complex test of ', 'quotes'] >>> line2argv('a more" complex test of " quotes') ['a', 'more complex test of ', 'quotes'] >>> line2argv('an "embedded \\"quote\\""') ['an', 'embedded "quote"'] # Komodo bug 48027 >>> line2argv('foo bar C:\\') ['foo', 'bar', 'C:\\'] # Komodo change 127581 >>> line2argv(r'"\test\slash" "foo bar" "foo\"bar"') ['\\test\\slash', 'foo bar', 'foo"bar'] # Komodo change 127629 >>> if sys.platform == "win32": ... line2argv(r'\foo\bar') == ['\\foo\\bar'] ... line2argv(r'\\foo\\bar') == ['\\\\foo\\\\bar'] ... line2argv('"foo') == ['foo'] ... else: ... line2argv(r'\foo\bar') == ['foobar'] ... line2argv(r'\\foo\\bar') == ['\\foo\\bar'] ... try: ... line2argv('"foo') ... except ValueError, ex: ... "not terminated" in str(ex) True True True """ import string line = line.strip() argv = [] state = "default" arg = None # the current argument being parsed i = -1 while 1: i += 1 if i >= len(line): break ch = line[i] if ch == "\\" and i+1 < len(line): # escaped char always added to arg, regardless of state if arg is None: arg = "" if (sys.platform == "win32" or state in ("double-quoted", "single-quoted") ) and line[i+1] not in tuple('"\''): arg += ch i += 1 arg += line[i] continue if state == "single-quoted": if ch == "'": state = "default" else: arg += ch elif state == "double-quoted": if ch == '"': state = "default" else: arg += ch elif state == "default": if ch == '"': if arg is None: arg = "" state = "double-quoted" elif ch == "'": if arg is None: arg = "" state = "single-quoted" elif ch in string.whitespace: if arg is not None: argv.append(arg) arg = None else: if arg is None: arg = "" arg += ch if arg is not None: argv.append(arg) if not sys.platform == "win32" and state != "default": raise ValueError("command line is not terminated: unfinished %s " "segment" % state) return argv def argv2line(argv): r"""Put together the given argument vector into a command line. "argv" is the argument vector to process. >>> from cmdln import argv2line >>> argv2line(['foo']) 'foo' >>> argv2line(['foo', 'bar']) 'foo bar' >>> argv2line(['foo', 'bar baz']) 'foo "bar baz"' >>> argv2line(['foo"bar']) 'foo"bar' >>> print argv2line(['foo" bar']) 'foo" bar' >>> print argv2line(["foo' bar"]) "foo' bar" >>> argv2line(["foo'bar"]) "foo'bar" """ escapedArgs = [] for arg in argv: if ' ' in arg and '"' not in arg: arg = '"'+arg+'"' elif ' ' in arg and "'" not in arg: arg = "'"+arg+"'" elif ' ' in arg: arg = arg.replace('"', r'\"') arg = '"'+arg+'"' escapedArgs.append(arg) return ' '.join(escapedArgs) # Recipe: dedent (0.1) in /Users/trentm/tm/recipes/cookbook def _dedentlines(lines, tabsize=8, skip_first_line=False): """_dedentlines(lines, tabsize=8, skip_first_line=False) -> dedented lines "lines" is a list of lines to dedent. "tabsize" is the tab width to use for indent width calculations. "skip_first_line" is a boolean indicating if the first line should be skipped for calculating the indent width and for dedenting. This is sometimes useful for docstrings and similar. Same as dedent() except operates on a sequence of lines. Note: the lines list is modified **in-place**. """ DEBUG = False if DEBUG: print "dedent: dedent(..., tabsize=%d, skip_first_line=%r)"\ % (tabsize, skip_first_line) indents = [] margin = None for i, line in enumerate(lines): if i == 0 and skip_first_line: continue indent = 0 for ch in line: if ch == ' ': indent += 1 elif ch == '\t': indent += tabsize - (indent % tabsize) elif ch in '\r\n': continue # skip all-whitespace lines else: break else: continue # skip all-whitespace lines if DEBUG: print "dedent: indent=%d: %r" % (indent, line) if margin is None: margin = indent else: margin = min(margin, indent) if DEBUG: print "dedent: margin=%r" % margin if margin is not None and margin > 0: for i, line in enumerate(lines): if i == 0 and skip_first_line: continue removed = 0 for j, ch in enumerate(line): if ch == ' ': removed += 1 elif ch == '\t': removed += tabsize - (removed % tabsize) elif ch in '\r\n': if DEBUG: print "dedent: %r: EOL -> strip up to EOL" % line lines[i] = lines[i][j:] break else: raise ValueError("unexpected non-whitespace char %r in " "line %r while removing %d-space margin" % (ch, line, margin)) if DEBUG: print "dedent: %r: %r -> removed %d/%d"\ % (line, ch, removed, margin) if removed == margin: lines[i] = lines[i][j+1:] break elif removed > margin: lines[i] = ' '*(removed-margin) + lines[i][j+1:] break return lines def _dedent(text, tabsize=8, skip_first_line=False): """_dedent(text, tabsize=8, skip_first_line=False) -> dedented text "text" is the text to dedent. "tabsize" is the tab width to use for indent width calculations. "skip_first_line" is a boolean indicating if the first line should be skipped for calculating the indent width and for dedenting. This is sometimes useful for docstrings and similar. textwrap.dedent(s), but don't expand tabs to spaces """ lines = text.splitlines(1) _dedentlines(lines, tabsize=tabsize, skip_first_line=skip_first_line) return ''.join(lines) def _get_indent(marker, s, tab_width=8): """_get_indent(marker, s, tab_width=8) -> (<indentation-of-'marker'>, <indentation-width>)""" # Figure out how much the marker is indented. INDENT_CHARS = tuple(' \t') start = s.index(marker) i = start while i > 0: if s[i-1] not in INDENT_CHARS: break i -= 1 indent = s[i:start] indent_width = 0 for ch in indent: if ch == ' ': indent_width += 1 elif ch == '\t': indent_width += tab_width - (indent_width % tab_width) return indent, indent_width def _get_trailing_whitespace(marker, s): """Return the whitespace content trailing the given 'marker' in string 's', up to and including a newline. """ suffix = '' start = s.index(marker) + len(marker) i = start while i < len(s): if s[i] in ' \t': suffix += s[i] elif s[i] in '\r\n': suffix += s[i] if s[i] == '\r' and i+1 < len(s) and s[i+1] == '\n': suffix += s[i+1] break else: break i += 1 return suffix #---- bash completion support # Note: This is still experimental. I expect to change this # significantly. # # To get Bash completion for a cmdln.Cmdln class, run the following # bash command: # $ complete -C 'python -m cmdln /path/to/script.py CmdlnClass' cmdname # For example: # $ complete -C 'python -m cmdln ~/bin/svn.py SVN' svn # #TODO: Simplify the above so don't have to given path to script (try to # find it on PATH, if possible). Could also make class name # optional if there is only one in the module (common case). if __name__ == "__main__" and len(sys.argv) == 6: def _log(s): return # no-op, comment out for debugging from os.path import expanduser fout = open(expanduser("~/tmp/bashcpln.log"), 'a') fout.write(str(s) + '\n') fout.close() # Recipe: module_from_path (1.0.1+) def _module_from_path(path): import imp, os, sys path = os.path.expanduser(path) dir = os.path.dirname(path) or os.curdir name = os.path.splitext(os.path.basename(path))[0] sys.path.insert(0, dir) try: iinfo = imp.find_module(name, [dir]) return imp.load_module(name, *iinfo) finally: sys.path.remove(dir) def _get_bash_cplns(script_path, class_name, cmd_name, token, preceding_token): _log('--') _log('get_cplns(%r, %r, %r, %r, %r)' % (script_path, class_name, cmd_name, token, preceding_token)) comp_line = os.environ["COMP_LINE"] comp_point = int(os.environ["COMP_POINT"]) _log("COMP_LINE: %r" % comp_line) _log("COMP_POINT: %r" % comp_point) try: script = _module_from_path(script_path) except ImportError, ex: _log("error importing `%s': %s" % (script_path, ex)) return [] shell = getattr(script, class_name)() cmd_map = shell._get_canonical_map() del cmd_map["EOF"] # Determine if completing the sub-command name. parts = comp_line[:comp_point].split(None, 1) _log(parts) if len(parts) == 1 or not (' ' in parts[1] or '\t' in parts[1]): #TODO: if parts[1].startswith('-'): handle top-level opts _log("complete sub-command names") matches = {} for name, canon_name in cmd_map.items(): if name.startswith(token): matches[name] = canon_name if not matches: return [] elif len(matches) == 1: return matches.keys() elif len(set(matches.values())) == 1: return [matches.values()[0]] else: return matches.keys() # Otherwise, complete options for the given sub-command. #TODO: refine this so it does the right thing with option args if token.startswith('-'): cmd_name = comp_line.split(None, 2)[1] try: cmd_canon_name = cmd_map[cmd_name] except KeyError: return [] handler = shell._get_cmd_handler(cmd_canon_name) optparser = getattr(handler, "optparser", None) if optparser is None: optparser = SubCmdOptionParser() opt_strs = [] for option in optparser.option_list: for opt_str in option._short_opts + option._long_opts: if opt_str.startswith(token): opt_strs.append(opt_str) return opt_strs return [] for cpln in _get_bash_cplns(*sys.argv[1:]): print cpln ########NEW FILE######## __FILENAME__ = errors #!/usr/bin/python -tt # # Copyright (c) 2007 Red Hat, Inc. # Copyright (c) 2011 Intel, Inc. # # This program is free software; you can redistribute it and/or modify it # under the terms of the GNU General Public License as published by the Free # Software Foundation; version 2 of the License # # This program is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY # or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License # for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., 59 # Temple Place - Suite 330, Boston, MA 02111-1307, USA. class CreatorError(Exception): """An exception base class for all imgcreate errors.""" keyword = '<creator>' def __init__(self, msg): self.msg = msg def __str__(self): if isinstance(self.msg, unicode): self.msg = self.msg.encode('utf-8', 'ignore') else: self.msg = str(self.msg) return self.keyword + self.msg class Usage(CreatorError): keyword = '<usage>' def __str__(self): if isinstance(self.msg, unicode): self.msg = self.msg.encode('utf-8', 'ignore') else: self.msg = str(self.msg) return self.keyword + self.msg + ', please use "--help" for more info' class Abort(CreatorError): keyword = '' class ConfigError(CreatorError): keyword = '<config>' class KsError(CreatorError): keyword = '<kickstart>' class RepoError(CreatorError): keyword = '<repo>' class RpmError(CreatorError): keyword = '<rpm>' class MountError(CreatorError): keyword = '<mount>' class SnapshotError(CreatorError): keyword = '<snapshot>' class SquashfsError(CreatorError): keyword = '<squashfs>' class BootstrapError(CreatorError): keyword = '<bootstrap>' class RuntimeError(CreatorError): keyword = '<runtime>' ########NEW FILE######## __FILENAME__ = Fiemap """ This module implements python API for the FIEMAP ioctl. The FIEMAP ioctl allows to find holes and mapped areas in a file. """ # Note, a lot of code in this module is not very readable, because it deals # with the rather complex FIEMAP ioctl. To understand the code, you need to # know the FIEMAP interface, which is documented in the # Documentation/filesystems/fiemap.txt file in the Linux kernel sources. # Disable the following pylint recommendations: # * Too many instance attributes (R0902) # pylint: disable=R0902 import os import struct import array import fcntl from mic.utils.misc import get_block_size # Format string for 'struct fiemap' _FIEMAP_FORMAT = "=QQLLLL" # sizeof(struct fiemap) _FIEMAP_SIZE = struct.calcsize(_FIEMAP_FORMAT) # Format string for 'struct fiemap_extent' _FIEMAP_EXTENT_FORMAT = "=QQQQQLLLL" # sizeof(struct fiemap_extent) _FIEMAP_EXTENT_SIZE = struct.calcsize(_FIEMAP_EXTENT_FORMAT) # The FIEMAP ioctl number _FIEMAP_IOCTL = 0xC020660B # Minimum buffer which is required for 'class Fiemap' to operate MIN_BUFFER_SIZE = _FIEMAP_SIZE + _FIEMAP_EXTENT_SIZE # The default buffer size for 'class Fiemap' DEFAULT_BUFFER_SIZE = 256 * 1024 class Error(Exception): """ A class for exceptions generated by this module. We currently support only one type of exceptions, and we basically throw human-readable problem description in case of errors. """ pass class Fiemap: """ This class provides API to the FIEMAP ioctl. Namely, it allows to iterate over all mapped blocks and over all holes. """ def _open_image_file(self): """ Open the image file. """ try: self._f_image = open(self._image_path, 'rb') except IOError as err: raise Error("cannot open image file '%s': %s" \ % (self._image_path, err)) self._f_image_needs_close = True def __init__(self, image, buf_size = DEFAULT_BUFFER_SIZE): """ Initialize a class instance. The 'image' argument is full path to the file to operate on, or a file object to operate on. The 'buf_size' argument is the size of the buffer for 'struct fiemap_extent' elements which will be used when invoking the FIEMAP ioctl. The larger is the buffer, the less times the FIEMAP ioctl will be invoked. """ self._f_image_needs_close = False if hasattr(image, "fileno"): self._f_image = image self._image_path = image.name else: self._image_path = image self._open_image_file() # Validate 'buf_size' if buf_size < MIN_BUFFER_SIZE: raise Error("too small buffer (%d bytes), minimum is %d bytes" \ % (buf_size, MIN_BUFFER_SIZE)) # How many 'struct fiemap_extent' elements fit the buffer buf_size -= _FIEMAP_SIZE self._fiemap_extent_cnt = buf_size / _FIEMAP_EXTENT_SIZE self._buf_size = self._fiemap_extent_cnt * _FIEMAP_EXTENT_SIZE self._buf_size += _FIEMAP_SIZE # Allocate a mutable buffer for the FIEMAP ioctl self._buf = array.array('B', [0] * self._buf_size) self.image_size = os.fstat(self._f_image.fileno()).st_size try: self.block_size = get_block_size(self._f_image) except IOError as err: raise Error("cannot get block size for '%s': %s" \ % (self._image_path, err)) self.blocks_cnt = self.image_size + self.block_size - 1 self.blocks_cnt /= self.block_size # Synchronize the image file to make sure FIEMAP returns correct values try: self._f_image.flush() except IOError as err: raise Error("cannot flush image file '%s': %s" \ % (self._image_path, err)) try: os.fsync(self._f_image.fileno()), except OSError as err: raise Error("cannot synchronize image file '%s': %s " \ % (self._image_path, err.strerror)) # Check if the FIEMAP ioctl is supported self.block_is_mapped(0) def __del__(self): """ The class destructor which closes the opened files. """ if self._f_image_needs_close: self._f_image.close() def _invoke_fiemap(self, block, count): """ Invoke the FIEMAP ioctl for 'count' blocks of the file starting from block number 'block'. The full result of the operation is stored in 'self._buf' on exit. Returns the unpacked 'struct fiemap' data structure in form of a python list (just like 'struct.upack()'). """ if block < 0 or block >= self.blocks_cnt: raise Error("bad block number %d, should be within [0, %d]" \ % (block, self.blocks_cnt)) # Initialize the 'struct fiemap' part of the buffer struct.pack_into(_FIEMAP_FORMAT, self._buf, 0, block * self.block_size, count * self.block_size, 0, 0, self._fiemap_extent_cnt, 0) try: fcntl.ioctl(self._f_image, _FIEMAP_IOCTL, self._buf, 1) except IOError as err: error_msg = "the FIEMAP ioctl failed for '%s': %s" \ % (self._image_path, err) if err.errno == os.errno.EPERM or err.errno == os.errno.EACCES: # The FIEMAP ioctl was added in kernel version 2.6.28 in 2008 error_msg += " (looks like your kernel does not support FIEMAP)" raise Error(error_msg) return struct.unpack(_FIEMAP_FORMAT, self._buf[:_FIEMAP_SIZE]) def block_is_mapped(self, block): """ This function returns 'True' if block number 'block' of the image file is mapped and 'False' otherwise. """ struct_fiemap = self._invoke_fiemap(block, 1) # The 3rd element of 'struct_fiemap' is the 'fm_mapped_extents' field. # If it contains zero, the block is not mapped, otherwise it is # mapped. return bool(struct_fiemap[3]) def block_is_unmapped(self, block): """ This function returns 'True' if block number 'block' of the image file is not mapped (hole) and 'False' otherwise. """ return not self.block_is_mapped(block) def _unpack_fiemap_extent(self, index): """ Unpack a 'struct fiemap_extent' structure object number 'index' from the internal 'self._buf' buffer. """ offset = _FIEMAP_SIZE + _FIEMAP_EXTENT_SIZE * index return struct.unpack(_FIEMAP_EXTENT_FORMAT, self._buf[offset : offset + _FIEMAP_EXTENT_SIZE]) def _do_get_mapped_ranges(self, start, count): """ Implements most the functionality for the 'get_mapped_ranges()' generator: invokes the FIEMAP ioctl, walks through the mapped extents and yields mapped block ranges. However, the ranges may be consecutive (e.g., (1, 100), (100, 200)) and 'get_mapped_ranges()' simply merges them. """ block = start while block < start + count: struct_fiemap = self._invoke_fiemap(block, count) mapped_extents = struct_fiemap[3] if mapped_extents == 0: # No more mapped blocks return extent = 0 while extent < mapped_extents: fiemap_extent = self._unpack_fiemap_extent(extent) # Start of the extent extent_start = fiemap_extent[0] # Starting block number of the extent extent_block = extent_start / self.block_size # Length of the extent extent_len = fiemap_extent[2] # Count of blocks in the extent extent_count = extent_len / self.block_size # Extent length and offset have to be block-aligned assert extent_start % self.block_size == 0 assert extent_len % self.block_size == 0 if extent_block > start + count - 1: return first = max(extent_block, block) last = min(extent_block + extent_count, start + count) - 1 yield (first, last) extent += 1 block = extent_block + extent_count def get_mapped_ranges(self, start, count): """ A generator which yields ranges of mapped blocks in the file. The ranges are tuples of 2 elements: [first, last], where 'first' is the first mapped block and 'last' is the last mapped block. The ranges are yielded for the area of the file of size 'count' blocks, starting from block 'start'. """ iterator = self._do_get_mapped_ranges(start, count) first_prev, last_prev = iterator.next() for first, last in iterator: if last_prev == first - 1: last_prev = last else: yield (first_prev, last_prev) first_prev, last_prev = first, last yield (first_prev, last_prev) def get_unmapped_ranges(self, start, count): """ Just like 'get_mapped_ranges()', but yields unmapped block ranges instead (holes). """ hole_first = start for first, last in self._do_get_mapped_ranges(start, count): if first > hole_first: yield (hole_first, first - 1) hole_first = last + 1 if hole_first < start + count: yield (hole_first, start + count - 1) ########NEW FILE######## __FILENAME__ = fs_related #!/usr/bin/python -tt # # Copyright (c) 2007, Red Hat, Inc. # Copyright (c) 2009, 2010, 2011 Intel, Inc. # # This program is free software; you can redistribute it and/or modify it # under the terms of the GNU General Public License as published by the Free # Software Foundation; version 2 of the License # # This program is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY # or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License # for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., 59 # Temple Place - Suite 330, Boston, MA 02111-1307, USA. from __future__ import with_statement import os import sys import errno import stat import random import string import time import uuid from mic import msger from mic.utils import runner from mic.utils.errors import * from mic.utils.oe.misc import * def find_binary_inchroot(binary, chroot): paths = ["/usr/sbin", "/usr/bin", "/sbin", "/bin" ] for path in paths: bin_path = "%s/%s" % (path, binary) if os.path.exists("%s/%s" % (chroot, bin_path)): return bin_path return None def find_binary_path(binary): if os.environ.has_key("PATH"): paths = os.environ["PATH"].split(":") else: paths = [] if os.environ.has_key("HOME"): paths += [os.environ["HOME"] + "/bin"] paths += ["/usr/local/sbin", "/usr/local/bin", "/usr/sbin", "/usr/bin", "/sbin", "/bin"] for path in paths: bin_path = "%s/%s" % (path, binary) if os.path.exists(bin_path): return bin_path print "External command '%s' not found, exiting." % binary print " (Please install '%s' on your host system)" % binary sys.exit(1) def makedirs(dirname): """A version of os.makedirs() that doesn't throw an exception if the leaf directory already exists. """ try: os.makedirs(dirname) except OSError, err: if err.errno != errno.EEXIST: raise def mksquashfs(in_img, out_img): fullpathmksquashfs = find_binary_path("mksquashfs") args = [fullpathmksquashfs, in_img, out_img] if not sys.stdout.isatty(): args.append("-no-progress") ret = runner.show(args) if ret != 0: raise SquashfsError("'%s' exited with error (%d)" % (' '.join(args), ret)) def resize2fs(fs, size): resize2fs = find_binary_path("resize2fs") if size == 0: # it means to minimalize it return runner.show([resize2fs, '-M', fs]) else: return runner.show([resize2fs, fs, "%sK" % (size / 1024,)]) def my_fuser(fp): fuser = find_binary_path("fuser") if not os.path.exists(fp): return False rc = runner.quiet([fuser, "-s", fp]) if rc == 0: for pid in runner.outs([fuser, fp]).split(): fd = open("/proc/%s/cmdline" % pid, "r") cmdline = fd.read() fd.close() if cmdline[:-1] == "/bin/bash": return True # not found return False class BindChrootMount: """Represents a bind mount of a directory into a chroot.""" def __init__(self, src, chroot, dest = None, option = None): self.root = os.path.abspath(os.path.expanduser(chroot)) self.option = option self.orig_src = self.src = src if os.path.islink(src): self.src = os.readlink(src) if not self.src.startswith('/'): self.src = os.path.abspath(os.path.join(os.path.dirname(src), self.src)) if not dest: dest = self.src self.dest = os.path.join(self.root, dest.lstrip('/')) self.mounted = False self.mountcmd = find_binary_path("mount") self.umountcmd = find_binary_path("umount") def ismounted(self): with open('/proc/mounts') as f: for line in f: if line.split()[1] == os.path.abspath(self.dest): return True return False def has_chroot_instance(self): lock = os.path.join(self.root, ".chroot.lock") return my_fuser(lock) def mount(self): if self.mounted or self.ismounted(): return makedirs(self.dest) rc = runner.show([self.mountcmd, "--bind", self.src, self.dest]) if rc != 0: raise MountError("Bind-mounting '%s' to '%s' failed" % (self.src, self.dest)) if self.option: rc = runner.show([self.mountcmd, "--bind", "-o", "remount,%s" % self.option, self.dest]) if rc != 0: raise MountError("Bind-remounting '%s' failed" % self.dest) self.mounted = True if os.path.islink(self.orig_src): dest = os.path.join(self.root, self.orig_src.lstrip('/')) if not os.path.exists(dest): os.symlink(self.src, dest) def unmount(self): if self.has_chroot_instance(): return if self.ismounted(): runner.show([self.umountcmd, "-l", self.dest]) self.mounted = False class LoopbackMount: """LoopbackMount compatibility layer for old API""" def __init__(self, lofile, mountdir, fstype = None): self.diskmount = DiskMount(LoopbackDisk(lofile,size = 0),mountdir,fstype,rmmountdir = True) self.losetup = False self.losetupcmd = find_binary_path("losetup") def cleanup(self): self.diskmount.cleanup() def unmount(self): self.diskmount.unmount() def lounsetup(self): if self.losetup: runner.show([self.losetupcmd, "-d", self.loopdev]) self.losetup = False self.loopdev = None def loopsetup(self): if self.losetup: return self.loopdev = get_loop_device(self.losetupcmd, self.lofile) self.losetup = True def mount(self): self.diskmount.mount() class SparseLoopbackMount(LoopbackMount): """SparseLoopbackMount compatibility layer for old API""" def __init__(self, lofile, mountdir, size, fstype = None): self.diskmount = DiskMount(SparseLoopbackDisk(lofile,size),mountdir,fstype,rmmountdir = True) def expand(self, create = False, size = None): self.diskmount.disk.expand(create, size) def truncate(self, size = None): self.diskmount.disk.truncate(size) def create(self): self.diskmount.disk.create() class SparseExtLoopbackMount(SparseLoopbackMount): """SparseExtLoopbackMount compatibility layer for old API""" def __init__(self, lofile, mountdir, size, fstype, blocksize, fslabel): self.diskmount = ExtDiskMount(SparseLoopbackDisk(lofile,size), mountdir, fstype, blocksize, fslabel, rmmountdir = True) def __format_filesystem(self): self.diskmount.__format_filesystem() def create(self): self.diskmount.disk.create() def resize(self, size = None): return self.diskmount.__resize_filesystem(size) def mount(self): self.diskmount.mount() def __fsck(self): self.extdiskmount.__fsck() def __get_size_from_filesystem(self): return self.diskmount.__get_size_from_filesystem() def __resize_to_minimal(self): return self.diskmount.__resize_to_minimal() def resparse(self, size = None): return self.diskmount.resparse(size) class Disk: """Generic base object for a disk The 'create' method must make the disk visible as a block device - eg by calling losetup. For RawDisk, this is obviously a no-op. The 'cleanup' method must undo the 'create' operation. """ def __init__(self, size, device = None): self._device = device self._size = size def create(self): pass def cleanup(self): pass def get_device(self): return self._device def set_device(self, path): self._device = path device = property(get_device, set_device) def get_size(self): return self._size size = property(get_size) class RawDisk(Disk): """A Disk backed by a block device. Note that create() is a no-op. """ def __init__(self, size, device): Disk.__init__(self, size, device) def fixed(self): return True def exists(self): return True class DiskImage(Disk): """ A Disk backed by a file. """ def __init__(self, image_file, size): Disk.__init__(self, size) self.image_file = image_file def exists(self): return os.path.exists(self.image_file) def create(self): if self.device is not None: return blocks = self.size / 1024 if self.size - blocks * 1024: blocks += 1 # create disk image dd_cmd = "dd if=/dev/zero of=%s bs=1024 seek=%d count=1" % \ (self.image_file, blocks) rc, out = exec_cmd(dd_cmd) self.device = self.image_file class LoopbackDisk(Disk): """A Disk backed by a file via the loop module.""" def __init__(self, lofile, size): Disk.__init__(self, size) self.lofile = lofile self.losetupcmd = find_binary_path("losetup") def fixed(self): return False def exists(self): return os.path.exists(self.lofile) def create(self): if self.device is not None: return self.device = get_loop_device(self.losetupcmd, self.lofile) def cleanup(self): if self.device is None: return msger.debug("Losetup remove %s" % self.device) rc = runner.show([self.losetupcmd, "-d", self.device]) self.device = None class SparseLoopbackDisk(LoopbackDisk): """A Disk backed by a sparse file via the loop module.""" def __init__(self, lofile, size): LoopbackDisk.__init__(self, lofile, size) def expand(self, create = False, size = None): flags = os.O_WRONLY if create: flags |= os.O_CREAT if not os.path.exists(self.lofile): makedirs(os.path.dirname(self.lofile)) if size is None: size = self.size msger.debug("Extending sparse file %s to %d" % (self.lofile, size)) if create: fd = os.open(self.lofile, flags, 0644) else: fd = os.open(self.lofile, flags) if size <= 0: size = 1 try: os.ftruncate(fd, size) except: # may be limited by 2G in 32bit env os.ftruncate(fd, 2**31L) os.close(fd) def truncate(self, size = None): if size is None: size = self.size msger.debug("Truncating sparse file %s to %d" % (self.lofile, size)) fd = os.open(self.lofile, os.O_WRONLY) os.ftruncate(fd, size) os.close(fd) def create(self): self.expand(create = True) LoopbackDisk.create(self) class Mount: """A generic base class to deal with mounting things.""" def __init__(self, mountdir): self.mountdir = mountdir def cleanup(self): self.unmount() def mount(self, options = None): pass def unmount(self): pass class DiskMount(Mount): """A Mount object that handles mounting of a Disk.""" def __init__(self, disk, mountdir, fstype = None, rmmountdir = True): Mount.__init__(self, mountdir) self.disk = disk self.fstype = fstype self.rmmountdir = rmmountdir self.mounted = False self.rmdir = False if fstype: self.mkfscmd = find_binary_path("mkfs." + self.fstype) else: self.mkfscmd = None self.mountcmd = find_binary_path("mount") self.umountcmd = find_binary_path("umount") def cleanup(self): Mount.cleanup(self) self.disk.cleanup() def unmount(self): if self.mounted: msger.debug("Unmounting directory %s" % self.mountdir) runner.quiet('sync') # sync the data on this mount point rc = runner.show([self.umountcmd, "-l", self.mountdir]) if rc == 0: self.mounted = False else: raise MountError("Failed to umount %s" % self.mountdir) if self.rmdir and not self.mounted: try: os.rmdir(self.mountdir) except OSError, e: pass self.rmdir = False def __create(self): self.disk.create() def mount(self, options = None): if self.mounted: return if not os.path.isdir(self.mountdir): msger.debug("Creating mount point %s" % self.mountdir) os.makedirs(self.mountdir) self.rmdir = self.rmmountdir self.__create() msger.debug("Mounting %s at %s" % (self.disk.device, self.mountdir)) if options: args = [ self.mountcmd, "-o", options, self.disk.device, self.mountdir ] else: args = [ self.mountcmd, self.disk.device, self.mountdir ] if self.fstype: args.extend(["-t", self.fstype]) rc = runner.show(args) if rc != 0: raise MountError("Failed to mount '%s' to '%s' with command '%s'. Retval: %s" % (self.disk.device, self.mountdir, " ".join(args), rc)) self.mounted = True class ExtDiskMount(DiskMount): """A DiskMount object that is able to format/resize ext[23] filesystems.""" def __init__(self, disk, mountdir, fstype, blocksize, fslabel, rmmountdir=True, skipformat = False, fsopts = None): DiskMount.__init__(self, disk, mountdir, fstype, rmmountdir) self.blocksize = blocksize self.fslabel = fslabel.replace("/", "") self.uuid = str(uuid.uuid4()) self.skipformat = skipformat self.fsopts = fsopts self.extopts = None self.dumpe2fs = find_binary_path("dumpe2fs") self.tune2fs = find_binary_path("tune2fs") def __parse_field(self, output, field): for line in output.split("\n"): if line.startswith(field + ":"): return line[len(field) + 1:].strip() raise KeyError("Failed to find field '%s' in output" % field) def __format_filesystem(self): if self.skipformat: msger.debug("Skip filesystem format.") return msger.verbose("Formating %s filesystem on %s" % (self.fstype, self.disk.device)) cmdlist = [self.mkfscmd, "-F", "-L", self.fslabel, "-m", "1", "-b", str(self.blocksize), "-U", self.uuid] if self.extopts: cmdlist.extend(self.extopts.split()) cmdlist.extend([self.disk.device]) rc, errout = runner.runtool(cmdlist, catch=2) if rc != 0: raise MountError("Error creating %s filesystem on disk %s:\n%s" % (self.fstype, self.disk.device, errout)) if not self.extopts: msger.debug("Tuning filesystem on %s" % self.disk.device) runner.show([self.tune2fs, "-c0", "-i0", "-Odir_index", "-ouser_xattr,acl", self.disk.device]) def __resize_filesystem(self, size = None): current_size = os.stat(self.disk.lofile)[stat.ST_SIZE] if size is None: size = self.disk.size if size == current_size: return if size > current_size: self.disk.expand(size) self.__fsck() resize2fs(self.disk.lofile, size) return size def __create(self): resize = False if not self.disk.fixed() and self.disk.exists(): resize = True self.disk.create() if resize: self.__resize_filesystem() else: self.__format_filesystem() def mount(self, options = None): self.__create() DiskMount.mount(self, options) def __fsck(self): msger.info("Checking filesystem %s" % self.disk.lofile) runner.quiet(["/sbin/e2fsck", "-f", "-y", self.disk.lofile]) def __get_size_from_filesystem(self): return int(self.__parse_field(runner.outs([self.dumpe2fs, '-h', self.disk.lofile]), "Block count")) * self.blocksize def __resize_to_minimal(self): self.__fsck() # # Use a binary search to find the minimal size # we can resize the image to # bot = 0 top = self.__get_size_from_filesystem() while top != (bot + 1): t = bot + ((top - bot) / 2) if not resize2fs(self.disk.lofile, t): top = t else: bot = t return top def resparse(self, size = None): self.cleanup() if size == 0: minsize = 0 else: minsize = self.__resize_to_minimal() self.disk.truncate(minsize) self.__resize_filesystem(size) return minsize class VfatDiskMount(DiskMount): """A DiskMount object that is able to format vfat/msdos filesystems.""" def __init__(self, disk, mountdir, fstype, blocksize, fslabel, rmmountdir=True, skipformat = False, fsopts = None): DiskMount.__init__(self, disk, mountdir, fstype, rmmountdir) self.blocksize = blocksize self.fslabel = fslabel.replace("/", "") rand1 = random.randint(0, 2**16 - 1) rand2 = random.randint(0, 2**16 - 1) self.uuid = "%04X-%04X" % (rand1, rand2) self.skipformat = skipformat self.fsopts = fsopts self.fsckcmd = find_binary_path("fsck." + self.fstype) def __format_filesystem(self): if self.skipformat: msger.debug("Skip filesystem format.") return msger.verbose("Formating %s filesystem on %s" % (self.fstype, self.disk.device)) rc = runner.show([self.mkfscmd, "-n", self.fslabel, "-i", self.uuid.replace("-", ""), self.disk.device]) if rc != 0: raise MountError("Error creating %s filesystem on disk %s" % (self.fstype,self.disk.device)) msger.verbose("Tuning filesystem on %s" % self.disk.device) def __resize_filesystem(self, size = None): current_size = os.stat(self.disk.lofile)[stat.ST_SIZE] if size is None: size = self.disk.size if size == current_size: return if size > current_size: self.disk.expand(size) self.__fsck() #resize2fs(self.disk.lofile, size) return size def __create(self): resize = False if not self.disk.fixed() and self.disk.exists(): resize = True self.disk.create() if resize: self.__resize_filesystem() else: self.__format_filesystem() def mount(self, options = None): self.__create() DiskMount.mount(self, options) def __fsck(self): msger.debug("Checking filesystem %s" % self.disk.lofile) runner.show([self.fsckcmd, "-y", self.disk.lofile]) def __get_size_from_filesystem(self): return self.disk.size def __resize_to_minimal(self): self.__fsck() # # Use a binary search to find the minimal size # we can resize the image to # bot = 0 top = self.__get_size_from_filesystem() return top def resparse(self, size = None): self.cleanup() minsize = self.__resize_to_minimal() self.disk.truncate(minsize) self.__resize_filesystem(size) return minsize class BtrfsDiskMount(DiskMount): """A DiskMount object that is able to format/resize btrfs filesystems.""" def __init__(self, disk, mountdir, fstype, blocksize, fslabel, rmmountdir=True, skipformat = False, fsopts = None): self.__check_btrfs() DiskMount.__init__(self, disk, mountdir, fstype, rmmountdir) self.blocksize = blocksize self.fslabel = fslabel.replace("/", "") self.uuid = None self.skipformat = skipformat self.fsopts = fsopts self.blkidcmd = find_binary_path("blkid") self.btrfsckcmd = find_binary_path("btrfsck") def __check_btrfs(self): found = False """ Need to load btrfs module to mount it """ load_module("btrfs") for line in open("/proc/filesystems").xreadlines(): if line.find("btrfs") > -1: found = True break if not found: raise MountError("Your system can't mount btrfs filesystem, please make sure your kernel has btrfs support and the module btrfs.ko has been loaded.") # disable selinux, selinux will block write if os.path.exists("/usr/sbin/setenforce"): runner.show(["/usr/sbin/setenforce", "0"]) def __parse_field(self, output, field): for line in output.split(" "): if line.startswith(field + "="): return line[len(field) + 1:].strip().replace("\"", "") raise KeyError("Failed to find field '%s' in output" % field) def __format_filesystem(self): if self.skipformat: msger.debug("Skip filesystem format.") return msger.verbose("Formating %s filesystem on %s" % (self.fstype, self.disk.device)) rc = runner.show([self.mkfscmd, "-L", self.fslabel, self.disk.device]) if rc != 0: raise MountError("Error creating %s filesystem on disk %s" % (self.fstype,self.disk.device)) self.uuid = self.__parse_field(runner.outs([self.blkidcmd, self.disk.device]), "UUID") def __resize_filesystem(self, size = None): current_size = os.stat(self.disk.lofile)[stat.ST_SIZE] if size is None: size = self.disk.size if size == current_size: return if size > current_size: self.disk.expand(size) self.__fsck() return size def __create(self): resize = False if not self.disk.fixed() and self.disk.exists(): resize = True self.disk.create() if resize: self.__resize_filesystem() else: self.__format_filesystem() def mount(self, options = None): self.__create() DiskMount.mount(self, options) def __fsck(self): msger.debug("Checking filesystem %s" % self.disk.lofile) runner.quiet([self.btrfsckcmd, self.disk.lofile]) def __get_size_from_filesystem(self): return self.disk.size def __resize_to_minimal(self): self.__fsck() return self.__get_size_from_filesystem() def resparse(self, size = None): self.cleanup() minsize = self.__resize_to_minimal() self.disk.truncate(minsize) self.__resize_filesystem(size) return minsize class DeviceMapperSnapshot(object): def __init__(self, imgloop, cowloop): self.imgloop = imgloop self.cowloop = cowloop self.__created = False self.__name = None self.dmsetupcmd = find_binary_path("dmsetup") """Load dm_snapshot if it isn't loaded""" load_module("dm_snapshot") def get_path(self): if self.__name is None: return None return os.path.join("/dev/mapper", self.__name) path = property(get_path) def create(self): if self.__created: return self.imgloop.create() self.cowloop.create() self.__name = "imgcreate-%d-%d" % (os.getpid(), random.randint(0, 2**16)) size = os.stat(self.imgloop.lofile)[stat.ST_SIZE] table = "0 %d snapshot %s %s p 8" % (size / 512, self.imgloop.device, self.cowloop.device) args = [self.dmsetupcmd, "create", self.__name, "--table", table] if runner.show(args) != 0: self.cowloop.cleanup() self.imgloop.cleanup() raise SnapshotError("Could not create snapshot device using: " + ' '.join(args)) self.__created = True def remove(self, ignore_errors = False): if not self.__created: return time.sleep(2) rc = runner.show([self.dmsetupcmd, "remove", self.__name]) if not ignore_errors and rc != 0: raise SnapshotError("Could not remove snapshot device") self.__name = None self.__created = False self.cowloop.cleanup() self.imgloop.cleanup() def get_cow_used(self): if not self.__created: return 0 # # dmsetup status on a snapshot returns e.g. # "0 8388608 snapshot 416/1048576" # or, more generally: # "A B snapshot C/D" # where C is the number of 512 byte sectors in use # out = runner.outs([self.dmsetupcmd, "status", self.__name]) try: return int((out.split()[3]).split('/')[0]) * 512 except ValueError: raise SnapshotError("Failed to parse dmsetup status: " + out) def create_image_minimizer(path, image, minimal_size): """ Builds a copy-on-write image which can be used to create a device-mapper snapshot of an image where the image's filesystem is as small as possible The steps taken are: 1) Create a sparse COW 2) Loopback mount the image and the COW 3) Create a device-mapper snapshot of the image using the COW 4) Resize the filesystem to the minimal size 5) Determine the amount of space used in the COW 6) Restroy the device-mapper snapshot 7) Truncate the COW, removing unused space 8) Create a squashfs of the COW """ imgloop = LoopbackDisk(image, None) # Passing bogus size - doesn't matter cowloop = SparseLoopbackDisk(os.path.join(os.path.dirname(path), "osmin"), 64L * 1024L * 1024L) snapshot = DeviceMapperSnapshot(imgloop, cowloop) try: snapshot.create() resize2fs(snapshot.path, minimal_size) cow_used = snapshot.get_cow_used() finally: snapshot.remove(ignore_errors = (not sys.exc_info()[0] is None)) cowloop.truncate(cow_used) mksquashfs(cowloop.lofile, path) os.unlink(cowloop.lofile) def load_module(module): found = False for line in open('/proc/modules').xreadlines(): if line.startswith("%s " % module): found = True break if not found: msger.info("Loading %s..." % module) runner.quiet(['modprobe', module]) class LoopDevice(object): def __init__(self, loopid=None): self.device = None self.loopid = loopid self.created = False self.kpartxcmd = find_binary_path("kpartx") self.losetupcmd = find_binary_path("losetup") def register(self, device): self.device = device self.loopid = None self.created = True def reg_atexit(self): import atexit atexit.register(self.close) def _genloopid(self): import glob if not glob.glob("/dev/loop[0-9]*"): return 10 fint = lambda x: x[9:].isdigit() and int(x[9:]) or 0 maxid = 1 + max(filter(lambda x: x<100, map(fint, glob.glob("/dev/loop[0-9]*")))) if maxid < 10: maxid = 10 if maxid >= 100: raise return maxid def _kpseek(self, device): rc, out = runner.runtool([self.kpartxcmd, '-l', '-v', device]) if rc != 0: raise MountError("Can't query dm snapshot on %s" % device) for line in out.splitlines(): if line and line.startswith("loop"): return True return False def _loseek(self, device): import re rc, out = runner.runtool([self.losetupcmd, '-a']) if rc != 0: raise MountError("Failed to run 'losetup -a'") for line in out.splitlines(): m = re.match("([^:]+): .*", line) if m and m.group(1) == device: return True return False def create(self): if not self.created: if not self.loopid: self.loopid = self._genloopid() self.device = "/dev/loop%d" % self.loopid if os.path.exists(self.device): if self._loseek(self.device): raise MountError("Device busy: %s" % self.device) else: self.created = True return mknod = find_binary_path('mknod') rc = runner.show([mknod, '-m664', self.device, 'b', '7', str(self.loopid)]) if rc != 0: raise MountError("Failed to create device %s" % self.device) else: self.created = True def close(self): if self.created: try: self.cleanup() self.device = None except MountError, e: msger.error("%s" % e) def cleanup(self): if self.device is None: return if self._kpseek(self.device): if self.created: for i in range(3, os.sysconf("SC_OPEN_MAX")): try: os.close(i) except: pass runner.quiet([self.kpartxcmd, "-d", self.device]) if self._loseek(self.device): runner.quiet([self.losetupcmd, "-d", self.device]) # FIXME: should sleep a while between two loseek if self._loseek(self.device): msger.warning("Can't cleanup loop device %s" % self.device) elif self.loopid: os.unlink(self.device) DEVICE_PIDFILE_DIR = "/var/tmp/mic/device" DEVICE_LOCKFILE = "/var/lock/__mic_loopdev.lock" def get_loop_device(losetupcmd, lofile): global DEVICE_PIDFILE_DIR global DEVICE_LOCKFILE import fcntl makedirs(os.path.dirname(DEVICE_LOCKFILE)) fp = open(DEVICE_LOCKFILE, 'w') fcntl.flock(fp, fcntl.LOCK_EX) try: loopdev = None devinst = LoopDevice() # clean up left loop device first clean_loop_devices() # provide an avaible loop device rc, out = runner.runtool([losetupcmd, "--find"]) if rc == 0: loopdev = out.split()[0] devinst.register(loopdev) if not loopdev or not os.path.exists(loopdev): devinst.create() loopdev = devinst.device # setup a loop device for image file rc = runner.show([losetupcmd, loopdev, lofile]) if rc != 0: raise MountError("Failed to setup loop device for '%s'" % lofile) devinst.reg_atexit() # try to save device and pid makedirs(DEVICE_PIDFILE_DIR) pidfile = os.path.join(DEVICE_PIDFILE_DIR, os.path.basename(loopdev)) if os.path.exists(pidfile): os.unlink(pidfile) with open(pidfile, 'w') as wf: wf.write(str(os.getpid())) except MountError, err: raise CreatorError("%s" % str(err)) except: raise finally: try: fcntl.flock(fp, fcntl.LOCK_UN) fp.close() os.unlink(DEVICE_LOCKFILE) except: pass return loopdev def clean_loop_devices(piddir=DEVICE_PIDFILE_DIR): if not os.path.exists(piddir) or not os.path.isdir(piddir): return for loopdev in os.listdir(piddir): pidfile = os.path.join(piddir, loopdev) try: with open(pidfile, 'r') as rf: devpid = int(rf.read()) except: devpid = None # if the process using this device is alive, skip it if not devpid or os.path.exists(os.path.join('/proc', str(devpid))): continue # try to clean it up try: devinst = LoopDevice() devinst.register(os.path.join('/dev', loopdev)) devinst.cleanup() os.unlink(pidfile) except: pass ########NEW FILE######## __FILENAME__ = gpt_parser #!/usr/bin/python -tt # # Copyright (c) 2013 Intel, Inc. # # This program is free software; you can redistribute it and/or modify it # under the terms of the GNU General Public License as published by the Free # Software Foundation; version 2 of the License # # This program is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY # or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License # for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., 59 # Temple Place - Suite 330, Boston, MA 02111-1307, USA. """ This module implements a simple GPT partitions parser which can read the GPT header and the GPT partition table. """ import struct import uuid import binascii from mic.utils.errors import MountError _GPT_HEADER_FORMAT = "<8s4sIIIQQQQ16sQIII" _GPT_HEADER_SIZE = struct.calcsize(_GPT_HEADER_FORMAT) _GPT_ENTRY_FORMAT = "<16s16sQQQ72s" _GPT_ENTRY_SIZE = struct.calcsize(_GPT_ENTRY_FORMAT) _SUPPORTED_GPT_REVISION = '\x00\x00\x01\x00' def _stringify_uuid(binary_uuid): """ A small helper function to transform a binary UUID into a string format. """ uuid_str = str(uuid.UUID(bytes_le = binary_uuid)) return uuid_str.upper() def _calc_header_crc(raw_hdr): """ Calculate GPT header CRC32 checksum. The 'raw_hdr' parameter has to be a list or a tuple containing all the elements of the GPT header in a "raw" form, meaning that it should simply contain "unpacked" disk data. """ raw_hdr = list(raw_hdr) raw_hdr[3] = 0 raw_hdr = struct.pack(_GPT_HEADER_FORMAT, *raw_hdr) return binascii.crc32(raw_hdr) & 0xFFFFFFFF def _validate_header(raw_hdr): """ Validate the GPT header. The 'raw_hdr' parameter has to be a list or a tuple containing all the elements of the GPT header in a "raw" form, meaning that it should simply contain "unpacked" disk data. """ # Validate the signature if raw_hdr[0] != 'EFI PART': raise MountError("GPT partition table not found") # Validate the revision if raw_hdr[1] != _SUPPORTED_GPT_REVISION: raise MountError("Unsupported GPT revision '%s', supported revision " \ "is '%s'" % \ (binascii.hexlify(raw_hdr[1]), binascii.hexlify(_SUPPORTED_GPT_REVISION))) # Validate header size if raw_hdr[2] != _GPT_HEADER_SIZE: raise MountError("Bad GPT header size: %d bytes, expected %d" % \ (raw_hdr[2], _GPT_HEADER_SIZE)) crc = _calc_header_crc(raw_hdr) if raw_hdr[3] != crc: raise MountError("GPT header crc mismatch: %#x, should be %#x" % \ (crc, raw_hdr[3])) class GptParser: """ GPT partition table parser. Allows reading the GPT header and the partition table, as well as modifying the partition table records. """ def __init__(self, disk_path, sector_size = 512): """ The class constructor which accepts the following parameters: * disk_path - full path to the disk image or device node * sector_size - size of a disk sector in bytes """ self.sector_size = sector_size self.disk_path = disk_path try: self._disk_obj = open(disk_path, 'r+b') except IOError as err: raise MountError("Cannot open file '%s' for reading GPT " \ "partitions: %s" % (disk_path, err)) def __del__(self): """ The class destructor. """ self._disk_obj.close() def _read_disk(self, offset, size): """ A helper function which reads 'size' bytes from offset 'offset' of the disk and checks all the error conditions. """ self._disk_obj.seek(offset) try: data = self._disk_obj.read(size) except IOError as err: raise MountError("cannot read from '%s': %s" % \ (self.disk_path, err)) if len(data) != size: raise MountError("cannot read %d bytes from offset '%d' of '%s', " \ "read only %d bytes" % \ (size, offset, self.disk_path, len(data))) return data def _write_disk(self, offset, buf): """ A helper function which writes buffer 'buf' to offset 'offset' of the disk. This function takes care of unaligned writes and checks all the error conditions. """ # Since we may be dealing with a block device, we only can write in # 'self.sector_size' chunks. Find the aligned starting and ending # disk offsets to read. start = (offset / self.sector_size) * self.sector_size end = ((start + len(buf)) / self.sector_size + 1) * self.sector_size data = self._read_disk(start, end - start) off = offset - start data = data[:off] + buf + data[off + len(buf):] self._disk_obj.seek(start) try: self._disk_obj.write(data) except IOError as err: raise MountError("cannot write to '%s': %s" % (self.disk_path, err)) def read_header(self, primary = True): """ Read and verify the GPT header and return a dictionary containing the following elements: 'signature' : header signature 'revision' : header revision 'hdr_size' : header size in bytes 'hdr_crc' : header CRC32 'hdr_lba' : LBA of this header 'hdr_offs' : byte disk offset of this header 'backup_lba' : backup header LBA 'backup_offs' : byte disk offset of backup header 'first_lba' : first usable LBA for partitions 'first_offs' : first usable byte disk offset for partitions 'last_lba' : last usable LBA for partitions 'last_offs' : last usable byte disk offset for partitions 'disk_uuid' : UUID of the disk 'ptable_lba' : starting LBA of array of partition entries 'ptable_offs' : disk byte offset of the start of the partition table 'ptable_size' : partition table size in bytes 'entries_cnt' : number of available partition table entries 'entry_size' : size of a single partition entry 'ptable_crc' : CRC32 of the partition table 'primary' : a boolean, if 'True', this is the primary GPT header, if 'False' - the secondary 'primary_str' : contains string "primary" if this is the primary GPT header, and "backup" otherwise This dictionary corresponds to the GPT header format. Please, see the UEFI standard for the description of these fields. If the 'primary' parameter is 'True', the primary GPT header is read, otherwise the backup GPT header is read instead. """ # Read and validate the primary GPT header raw_hdr = self._read_disk(self.sector_size, _GPT_HEADER_SIZE) raw_hdr = struct.unpack(_GPT_HEADER_FORMAT, raw_hdr) _validate_header(raw_hdr) primary_str = "primary" if not primary: # Read and validate the backup GPT header raw_hdr = self._read_disk(raw_hdr[6] * self.sector_size, _GPT_HEADER_SIZE) raw_hdr = struct.unpack(_GPT_HEADER_FORMAT, raw_hdr) _validate_header(raw_hdr) primary_str = "backup" return { 'signature' : raw_hdr[0], 'revision' : raw_hdr[1], 'hdr_size' : raw_hdr[2], 'hdr_crc' : raw_hdr[3], 'hdr_lba' : raw_hdr[5], 'hdr_offs' : raw_hdr[5] * self.sector_size, 'backup_lba' : raw_hdr[6], 'backup_offs' : raw_hdr[6] * self.sector_size, 'first_lba' : raw_hdr[7], 'first_offs' : raw_hdr[7] * self.sector_size, 'last_lba' : raw_hdr[8], 'last_offs' : raw_hdr[8] * self.sector_size, 'disk_uuid' :_stringify_uuid(raw_hdr[9]), 'ptable_lba' : raw_hdr[10], 'ptable_offs' : raw_hdr[10] * self.sector_size, 'ptable_size' : raw_hdr[11] * raw_hdr[12], 'entries_cnt' : raw_hdr[11], 'entry_size' : raw_hdr[12], 'ptable_crc' : raw_hdr[13], 'primary' : primary, 'primary_str' : primary_str } def _read_raw_ptable(self, header): """ Read and validate primary or backup partition table. The 'header' argument is the GPT header. If it is the primary GPT header, then the primary partition table is read and validated, otherwise - the backup one. The 'header' argument is a dictionary which is returned by the 'read_header()' method. """ raw_ptable = self._read_disk(header['ptable_offs'], header['ptable_size']) crc = binascii.crc32(raw_ptable) & 0xFFFFFFFF if crc != header['ptable_crc']: raise MountError("Partition table at LBA %d (%s) is corrupted" % \ (header['ptable_lba'], header['primary_str'])) return raw_ptable def get_partitions(self, primary = True): """ This is a generator which parses the GPT partition table and generates the following dictionary for each partition: 'index' : the index of the partition table endry 'offs' : byte disk offset of the partition table entry 'type_uuid' : partition type UUID 'part_uuid' : partition UUID 'first_lba' : the first LBA 'last_lba' : the last LBA 'flags' : attribute flags 'name' : partition name 'primary' : a boolean, if 'True', this is the primary partition table, if 'False' - the secondary 'primary_str' : contains string "primary" if this is the primary GPT header, and "backup" otherwise This dictionary corresponds to the GPT header format. Please, see the UEFI standard for the description of these fields. If the 'primary' parameter is 'True', partitions from the primary GPT partition table are generated, otherwise partitions from the backup GPT partition table are generated. """ if primary: primary_str = "primary" else: primary_str = "backup" header = self.read_header(primary) raw_ptable = self._read_raw_ptable(header) for index in xrange(0, header['entries_cnt']): start = header['entry_size'] * index end = start + header['entry_size'] raw_entry = struct.unpack(_GPT_ENTRY_FORMAT, raw_ptable[start:end]) if raw_entry[2] == 0 or raw_entry[3] == 0: continue part_name = str(raw_entry[5].decode('UTF-16').split('\0', 1)[0]) yield { 'index' : index, 'offs' : header['ptable_offs'] + start, 'type_uuid' : _stringify_uuid(raw_entry[0]), 'part_uuid' : _stringify_uuid(raw_entry[1]), 'first_lba' : raw_entry[2], 'last_lba' : raw_entry[3], 'flags' : raw_entry[4], 'name' : part_name, 'primary' : primary, 'primary_str' : primary_str } def _change_partition(self, header, entry): """ A helper function for 'change_partitions()' which changes a a paricular instance of the partition table (primary or backup). """ if entry['index'] >= header['entries_cnt']: raise MountError("Partition table at LBA %d has only %d " \ "records cannot change record number %d" % \ (header['entries_cnt'], entry['index'])) # Read raw GPT header raw_hdr = self._read_disk(header['hdr_offs'], _GPT_HEADER_SIZE) raw_hdr = list(struct.unpack(_GPT_HEADER_FORMAT, raw_hdr)) _validate_header(raw_hdr) # Prepare the new partition table entry raw_entry = struct.pack(_GPT_ENTRY_FORMAT, uuid.UUID(entry['type_uuid']).bytes_le, uuid.UUID(entry['part_uuid']).bytes_le, entry['first_lba'], entry['last_lba'], entry['flags'], entry['name'].encode('UTF-16')) # Write the updated entry to the disk entry_offs = header['ptable_offs'] + \ header['entry_size'] * entry['index'] self._write_disk(entry_offs, raw_entry) # Calculate and update partition table CRC32 raw_ptable = self._read_disk(header['ptable_offs'], header['ptable_size']) raw_hdr[13] = binascii.crc32(raw_ptable) & 0xFFFFFFFF # Calculate and update the GPT header CRC raw_hdr[3] = _calc_header_crc(raw_hdr) # Write the updated header to the disk raw_hdr = struct.pack(_GPT_HEADER_FORMAT, *raw_hdr) self._write_disk(header['hdr_offs'], raw_hdr) def change_partition(self, entry): """ Change a GPT partition. The 'entry' argument has the same format as 'get_partitions()' returns. This function simply changes the partition table record corresponding to 'entry' in both, the primary and the backup GPT partition tables. The parition table CRC is re-calculated and the GPT headers are modified accordingly. """ # Change the primary partition table header = self.read_header(True) self._change_partition(header, entry) # Change the backup partition table header = self.read_header(False) self._change_partition(header, entry) ########NEW FILE######## __FILENAME__ = grabber #!/usr/bin/python import os import sys import rpm import fcntl import struct import termios from mic import msger from mic.utils import runner from mic.utils.errors import CreatorError from urlgrabber import grabber from urlgrabber import __version__ as grabber_version if rpm.labelCompare(grabber_version.split('.'), '3.9.0'.split('.')) == -1: msger.warning("Version of python-urlgrabber is %s, lower than '3.9.0', " "you may encounter some network issues" % grabber_version) def myurlgrab(url, filename, proxies, progress_obj = None): g = grabber.URLGrabber() if progress_obj is None: progress_obj = TextProgress() if url.startswith("file:/"): filepath = "/%s" % url.replace("file:", "").lstrip('/') if not os.path.exists(filepath): raise CreatorError("URLGrabber error: can't find file %s" % url) if url.endswith('.rpm'): return filepath else: # untouch repometadata in source path runner.show(['cp', '-f', filepath, filename]) else: try: filename = g.urlgrab(url=str(url), filename=filename, ssl_verify_host=False, ssl_verify_peer=False, proxies=proxies, http_headers=(('Pragma', 'no-cache'),), quote=0, progress_obj=progress_obj) except grabber.URLGrabError, err: msg = str(err) if msg.find(url) < 0: msg += ' on %s' % url raise CreatorError(msg) return filename def terminal_width(fd=1): """ Get the real terminal width """ try: buf = 'abcdefgh' buf = fcntl.ioctl(fd, termios.TIOCGWINSZ, buf) return struct.unpack('hhhh', buf)[1] except: # IOError return 80 def truncate_url(url, width): return os.path.basename(url)[0:width] class TextProgress(object): # make the class as singleton _instance = None def __new__(cls, *args, **kwargs): if not cls._instance: cls._instance = super(TextProgress, cls).__new__(cls, *args, **kwargs) return cls._instance def __init__(self, totalnum = None): self.total = totalnum self.counter = 1 def start(self, filename, url, *args, **kwargs): self.url = url self.termwidth = terminal_width() msger.info("\r%-*s" % (self.termwidth, " ")) if self.total is None: msger.info("\rRetrieving %s ..." % truncate_url(self.url, self.termwidth - 15)) else: msger.info("\rRetrieving %s [%d/%d] ..." % (truncate_url(self.url, self.termwidth - 25), self.counter, self.total)) def update(self, *args): pass def end(self, *args): if self.counter == self.total: msger.raw("\n") if self.total is not None: self.counter += 1 ########NEW FILE######## __FILENAME__ = misc #!/usr/bin/python -tt # # Copyright (c) 2010, 2011 Intel Inc. # # This program is free software; you can redistribute it and/or modify it # under the terms of the GNU General Public License as published by the Free # Software Foundation; version 2 of the License # # This program is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY # or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License # for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., 59 # Temple Place - Suite 330, Boston, MA 02111-1307, USA. import os import sys import time import tempfile import re import shutil import glob import hashlib import subprocess import platform import traceback try: import sqlite3 as sqlite except ImportError: import sqlite try: from xml.etree import cElementTree except ImportError: import cElementTree xmlparse = cElementTree.parse from mic import msger from mic.utils.errors import CreatorError, SquashfsError from mic.utils.fs_related import find_binary_path, makedirs from mic.utils.proxy import get_proxy_for from mic.utils import runner RPM_RE = re.compile("(.*)\.(.*) (.*)-(.*)") RPM_FMT = "%(name)s.%(arch)s %(version)s-%(release)s" SRPM_RE = re.compile("(.*)-(\d+.*)-(\d+\.\d+).src.rpm") def build_name(kscfg, release=None, prefix = None, suffix = None): """Construct and return an image name string. This is a utility function to help create sensible name and fslabel strings. The name is constructed using the sans-prefix-and-extension kickstart filename and the supplied prefix and suffix. kscfg -- a path to a kickstart file release -- a replacement to suffix for image release prefix -- a prefix to prepend to the name; defaults to None, which causes no prefix to be used suffix -- a suffix to append to the name; defaults to None, which causes a YYYYMMDDHHMM suffix to be used Note, if maxlen is less then the len(suffix), you get to keep both pieces. """ name = os.path.basename(kscfg) idx = name.rfind('.') if idx >= 0: name = name[:idx] if release is not None: suffix = "" if prefix is None: prefix = "" if suffix is None: suffix = time.strftime("%Y%m%d%H%M") if name.startswith(prefix): name = name[len(prefix):] prefix = "%s-" % prefix if prefix else "" suffix = "-%s" % suffix if suffix else "" ret = prefix + name + suffix return ret def get_distro(): """Detect linux distribution, support "meego" """ support_dists = ('SuSE', 'debian', 'fedora', 'redhat', 'centos', 'meego', 'moblin', 'tizen') try: (dist, ver, id) = platform.linux_distribution( \ supported_dists = support_dists) except: (dist, ver, id) = platform.dist( \ supported_dists = support_dists) return (dist, ver, id) def get_distro_str(): """Get composited string for current linux distribution """ (dist, ver, id) = get_distro() if not dist: return 'Unknown Linux Distro' else: distro_str = ' '.join(map(str.strip, (dist, ver, id))) return distro_str.strip() _LOOP_RULE_PTH = None def hide_loopdev_presentation(): udev_rules = "80-prevent-loop-present.rules" udev_rules_dir = [ '/usr/lib/udev/rules.d/', '/lib/udev/rules.d/', '/etc/udev/rules.d/' ] global _LOOP_RULE_PTH for rdir in udev_rules_dir: if os.path.exists(rdir): _LOOP_RULE_PTH = os.path.join(rdir, udev_rules) if not _LOOP_RULE_PTH: return try: with open(_LOOP_RULE_PTH, 'w') as wf: wf.write('KERNEL=="loop*", ENV{UDISKS_PRESENTATION_HIDE}="1"') runner.quiet('udevadm trigger') except: pass def unhide_loopdev_presentation(): global _LOOP_RULE_PTH if not _LOOP_RULE_PTH: return try: os.unlink(_LOOP_RULE_PTH) runner.quiet('udevadm trigger') except: pass def extract_rpm(rpmfile, targetdir): rpm2cpio = find_binary_path("rpm2cpio") cpio = find_binary_path("cpio") olddir = os.getcwd() os.chdir(targetdir) msger.verbose("Extract rpm file with cpio: %s" % rpmfile) p1 = subprocess.Popen([rpm2cpio, rpmfile], stdout=subprocess.PIPE) p2 = subprocess.Popen([cpio, "-idv"], stdin=p1.stdout, stdout=subprocess.PIPE, stderr=subprocess.PIPE) (sout, serr) = p2.communicate() msger.verbose(sout or serr) os.chdir(olddir) def compressing(fpath, method): comp_map = { "gz": "gzip", "bz2": "bzip2" } if method not in comp_map: raise CreatorError("Unsupport compress format: %s, valid values: %s" % (method, ','.join(comp_map.keys()))) cmd = find_binary_path(comp_map[method]) rc = runner.show([cmd, "-f", fpath]) if rc: raise CreatorError("Failed to %s file: %s" % (comp_map[method], fpath)) def taring(dstfile, target): import tarfile basen, ext = os.path.splitext(dstfile) comp = {".tar": None, ".gz": "gz", # for .tar.gz ".bz2": "bz2", # for .tar.bz2 ".tgz": "gz", ".tbz": "bz2"}[ext] # specify tarball file path if not comp: tarpath = dstfile elif basen.endswith(".tar"): tarpath = basen else: tarpath = basen + ".tar" wf = tarfile.open(tarpath, 'w') if os.path.isdir(target): for item in os.listdir(target): wf.add(os.path.join(target, item), item) else: wf.add(target, os.path.basename(target)) wf.close() if comp: compressing(tarpath, comp) # when dstfile ext is ".tgz" and ".tbz", should rename if not basen.endswith(".tar"): shutil.move("%s.%s" % (tarpath, comp), dstfile) def ziping(dstfile, target): import zipfile wf = zipfile.ZipFile(dstfile, 'w', compression=zipfile.ZIP_DEFLATED) if os.path.isdir(target): for item in os.listdir(target): fpath = os.path.join(target, item) if not os.path.isfile(fpath): continue wf.write(fpath, item, zipfile.ZIP_DEFLATED) else: wf.write(target, os.path.basename(target), zipfile.ZIP_DEFLATED) wf.close() pack_formats = { ".tar": taring, ".tar.gz": taring, ".tar.bz2": taring, ".tgz": taring, ".tbz": taring, ".zip": ziping, } def packing(dstfile, target): (base, ext) = os.path.splitext(dstfile) if ext in (".gz", ".bz2") and base.endswith(".tar"): ext = ".tar" + ext if ext not in pack_formats: raise CreatorError("Unsupport pack format: %s, valid values: %s" % (ext, ','.join(pack_formats.keys()))) func = pack_formats[ext] # func should be callable func(dstfile, target) def human_size(size): """Return human readable string for Bytes size """ if size <= 0: return "0M" import math measure = ['B', 'K', 'M', 'G', 'T', 'P', 'E', 'Z', 'Y'] expo = int(math.log(size, 1024)) mant = float(size/math.pow(1024, expo)) return "{0:.1f}{1:s}".format(mant, measure[expo]) def get_block_size(file_obj): """ Returns block size for file object 'file_obj'. Errors are indicated by the 'IOError' exception. """ from fcntl import ioctl import struct # Get the block size of the host file-system for the image file by calling # the FIGETBSZ ioctl (number 2). binary_data = ioctl(file_obj, 2, struct.pack('I', 0)) return struct.unpack('I', binary_data)[0] def check_space_pre_cp(src, dst): """Check whether disk space is enough before 'cp' like operations, else exception will be raised. """ srcsize = get_file_size(src) * 1024 * 1024 freesize = get_filesystem_avail(dst) if srcsize > freesize: raise CreatorError("space on %s(%s) is not enough for about %s files" % (dst, human_size(freesize), human_size(srcsize))) def calc_hashes(file_path, hash_names, start = 0, end = None): """ Calculate hashes for a file. The 'file_path' argument is the file to calculate hash functions for, 'start' and 'end' are the starting and ending file offset to calculate the has functions for. The 'hash_names' argument is a list of hash names to calculate. Returns the the list of calculated hash values in the hexadecimal form in the same order as 'hash_names'. """ if end == None: end = os.path.getsize(file_path) chunk_size = 65536 to_read = end - start read = 0 hashes = [] for hash_name in hash_names: hashes.append(hashlib.new(hash_name)) with open(file_path, "rb") as f: f.seek(start) while read < to_read: if read + chunk_size > to_read: chunk_size = to_read - read chunk = f.read(chunk_size) for hash_obj in hashes: hash_obj.update(chunk) read += chunk_size result = [] for hash_obj in hashes: result.append(hash_obj.hexdigest()) return result def get_md5sum(fpath): return calc_hashes(fpath, ('md5', ))[0] def normalize_ksfile(ksconf, release, arch): ''' Return the name of a normalized ks file in which macro variables @BUILD_ID@ and @ARCH@ are replace with real values. The original ks file is returned if no special macro is used, otherwise a temp file is created and returned, which will be deleted when program exits normally. ''' if not release: release = "latest" if not arch or re.match(r'i.86', arch): arch = "ia32" with open(ksconf) as f: ksc = f.read() if "@ARCH@" not in ksc and "@BUILD_ID@" not in ksc: return ksconf msger.info("Substitute macro variable @BUILD_ID@/@ARCH@ in ks: %s" % ksconf) ksc = ksc.replace("@ARCH@", arch) ksc = ksc.replace("@BUILD_ID@", release) fd, ksconf = tempfile.mkstemp(prefix=os.path.basename(ksconf)) os.write(fd, ksc) os.close(fd) msger.debug('normalized ks file:%s' % ksconf) def remove_temp_ks(): try: os.unlink(ksconf) except OSError, err: msger.warning('Failed to remove temp ks file:%s:%s' % (ksconf, err)) import atexit atexit.register(remove_temp_ks) return ksconf def _check_mic_chroot(rootdir): def _path(path): return rootdir.rstrip('/') + path release_files = map(_path, [ "/etc/moblin-release", "/etc/meego-release", "/etc/tizen-release"]) if not any(map(os.path.exists, release_files)): msger.warning("Dir %s is not a MeeGo/Tizen chroot env" % rootdir) if not glob.glob(rootdir + "/boot/vmlinuz-*"): msger.warning("Failed to find kernel module under %s" % rootdir) return def selinux_check(arch, fstypes): try: getenforce = find_binary_path('getenforce') except CreatorError: return selinux_status = runner.outs([getenforce]) if arch and arch.startswith("arm") and selinux_status == "Enforcing": raise CreatorError("Can't create arm image if selinux is enabled, " "please run 'setenforce 0' to disable selinux") use_btrfs = filter(lambda typ: typ == 'btrfs', fstypes) if use_btrfs and selinux_status == "Enforcing": raise CreatorError("Can't create btrfs image if selinux is enabled," " please run 'setenforce 0' to disable selinux") def get_image_type(path): def _get_extension_name(path): match = re.search("(?<=\.)\w+$", path) if match: return match.group(0) else: return None if os.path.isdir(path): _check_mic_chroot(path) return "fs" maptab = { "tar": "loop", "raw":"raw", "vmdk":"vmdk", "vdi":"vdi", "iso":"livecd", "usbimg":"liveusb", } extension = _get_extension_name(path) if extension in maptab: return maptab[extension] fd = open(path, "rb") file_header = fd.read(1024) fd.close() vdi_flag = "<<< Sun VirtualBox Disk Image >>>" if file_header[0:len(vdi_flag)] == vdi_flag: return maptab["vdi"] output = runner.outs(['file', path]) isoptn = re.compile(r".*ISO 9660 CD-ROM filesystem.*(bootable).*") usbimgptn = re.compile(r".*x86 boot sector.*active.*") rawptn = re.compile(r".*x86 boot sector.*") vmdkptn = re.compile(r".*VMware. disk image.*") ext3fsimgptn = re.compile(r".*Linux.*ext3 filesystem data.*") ext4fsimgptn = re.compile(r".*Linux.*ext4 filesystem data.*") btrfsimgptn = re.compile(r".*BTRFS.*") if isoptn.match(output): return maptab["iso"] elif usbimgptn.match(output): return maptab["usbimg"] elif rawptn.match(output): return maptab["raw"] elif vmdkptn.match(output): return maptab["vmdk"] elif ext3fsimgptn.match(output): return "ext3fsimg" elif ext4fsimgptn.match(output): return "ext4fsimg" elif btrfsimgptn.match(output): return "btrfsimg" else: raise CreatorError("Cannot detect the type of image: %s" % path) def get_file_size(filename): """ Return size in MB unit """ cmd = ['du', "-s", "-b", "-B", "1M", filename] rc, duOutput = runner.runtool(cmd) if rc != 0: raise CreatorError("Failed to run: %s" % ' '.join(cmd)) size1 = int(duOutput.split()[0]) cmd = ['du', "-s", "-B", "1M", filename] rc, duOutput = runner.runtool(cmd) if rc != 0: raise CreatorError("Failed to run: %s" % ' '.join(cmd)) size2 = int(duOutput.split()[0]) return max(size1, size2) def get_filesystem_avail(fs): vfstat = os.statvfs(fs) return vfstat.f_bavail * vfstat.f_bsize def convert_image(srcimg, srcfmt, dstimg, dstfmt): #convert disk format if dstfmt != "raw": raise CreatorError("Invalid destination image format: %s" % dstfmt) msger.debug("converting %s image to %s" % (srcimg, dstimg)) if srcfmt == "vmdk": path = find_binary_path("qemu-img") argv = [path, "convert", "-f", "vmdk", srcimg, "-O", dstfmt, dstimg] elif srcfmt == "vdi": path = find_binary_path("VBoxManage") argv = [path, "internalcommands", "converttoraw", srcimg, dstimg] else: raise CreatorError("Invalid soure image format: %s" % srcfmt) rc = runner.show(argv) if rc == 0: msger.debug("convert successful") if rc != 0: raise CreatorError("Unable to convert disk to %s" % dstfmt) def uncompress_squashfs(squashfsimg, outdir): """Uncompress file system from squshfs image""" unsquashfs = find_binary_path("unsquashfs") args = [ unsquashfs, "-d", outdir, squashfsimg ] rc = runner.show(args) if (rc != 0): raise SquashfsError("Failed to uncompress %s." % squashfsimg) def mkdtemp(dir = "/var/tmp", prefix = "wic-tmp-"): """ FIXME: use the dir in wic.conf instead """ makedirs(dir) return tempfile.mkdtemp(dir = dir, prefix = prefix) def get_repostrs_from_ks(ks): def _get_temp_reponame(baseurl): md5obj = hashlib.md5(baseurl) tmpreponame = "%s" % md5obj.hexdigest() return tmpreponame kickstart_repos = [] for repodata in ks.handler.repo.repoList: repo = {} for attr in ('name', 'baseurl', 'mirrorlist', 'includepkgs', # val is list 'excludepkgs', # val is list 'cost', # int 'priority',# int 'save', 'proxy', 'proxyuser', 'proxypasswd', 'proxypasswd', 'debuginfo', 'source', 'gpgkey', 'ssl_verify'): if hasattr(repodata, attr) and getattr(repodata, attr): repo[attr] = getattr(repodata, attr) if 'name' not in repo: repo['name'] = _get_temp_reponame(repodata.baseurl) kickstart_repos.append(repo) return kickstart_repos def _get_uncompressed_data_from_url(url, filename, proxies): filename = myurlgrab(url, filename, proxies) suffix = None if filename.endswith(".gz"): suffix = ".gz" runner.quiet(['gunzip', "-f", filename]) elif filename.endswith(".bz2"): suffix = ".bz2" runner.quiet(['bunzip2', "-f", filename]) if suffix: filename = filename.replace(suffix, "") return filename def _get_metadata_from_repo(baseurl, proxies, cachedir, reponame, filename, sumtype=None, checksum=None): url = os.path.join(baseurl, filename) filename_tmp = str("%s/%s/%s" % (cachedir, reponame, os.path.basename(filename))) if os.path.splitext(filename_tmp)[1] in (".gz", ".bz2"): filename = os.path.splitext(filename_tmp)[0] else: filename = filename_tmp if sumtype and checksum and os.path.exists(filename): try: sumcmd = find_binary_path("%ssum" % sumtype) except: file_checksum = None else: file_checksum = runner.outs([sumcmd, filename]).split()[0] if file_checksum and file_checksum == checksum: return filename return _get_uncompressed_data_from_url(url,filename_tmp,proxies) def get_metadata_from_repos(repos, cachedir): my_repo_metadata = [] for repo in repos: reponame = repo['name'] baseurl = repo['baseurl'] if 'proxy' in repo: proxy = repo['proxy'] else: proxy = get_proxy_for(baseurl) proxies = None if proxy: proxies = {str(baseurl.split(":")[0]):str(proxy)} makedirs(os.path.join(cachedir, reponame)) url = os.path.join(baseurl, "repodata/repomd.xml") filename = os.path.join(cachedir, reponame, 'repomd.xml') repomd = myurlgrab(url, filename, proxies) try: root = xmlparse(repomd) except SyntaxError: raise CreatorError("repomd.xml syntax error.") ns = root.getroot().tag ns = ns[0:ns.rindex("}")+1] filepaths = {} checksums = {} sumtypes = {} for elm in root.getiterator("%sdata" % ns): if elm.attrib["type"] == "patterns": filepaths['patterns'] = elm.find("%slocation" % ns).attrib['href'] checksums['patterns'] = elm.find("%sopen-checksum" % ns).text sumtypes['patterns'] = elm.find("%sopen-checksum" % ns).attrib['type'] break for elm in root.getiterator("%sdata" % ns): if elm.attrib["type"] in ("group_gz", "group"): filepaths['comps'] = elm.find("%slocation" % ns).attrib['href'] checksums['comps'] = elm.find("%sopen-checksum" % ns).text sumtypes['comps'] = elm.find("%sopen-checksum" % ns).attrib['type'] break primary_type = None for elm in root.getiterator("%sdata" % ns): if elm.attrib["type"] in ("primary_db", "primary"): primary_type = elm.attrib["type"] filepaths['primary'] = elm.find("%slocation" % ns).attrib['href'] checksums['primary'] = elm.find("%sopen-checksum" % ns).text sumtypes['primary'] = elm.find("%sopen-checksum" % ns).attrib['type'] break if not primary_type: continue for item in ("primary", "patterns", "comps"): if item not in filepaths: filepaths[item] = None continue if not filepaths[item]: continue filepaths[item] = _get_metadata_from_repo(baseurl, proxies, cachedir, reponame, filepaths[item], sumtypes[item], checksums[item]) """ Get repo key """ try: repokey = _get_metadata_from_repo(baseurl, proxies, cachedir, reponame, "repodata/repomd.xml.key") except CreatorError: repokey = None msger.debug("\ncan't get %s/%s" % (baseurl, "repodata/repomd.xml.key")) my_repo_metadata.append({"name":reponame, "baseurl":baseurl, "repomd":repomd, "primary":filepaths['primary'], "cachedir":cachedir, "proxies":proxies, "patterns":filepaths['patterns'], "comps":filepaths['comps'], "repokey":repokey}) return my_repo_metadata def get_rpmver_in_repo(repometadata): for repo in repometadata: if repo["primary"].endswith(".xml"): root = xmlparse(repo["primary"]) ns = root.getroot().tag ns = ns[0:ns.rindex("}")+1] versionlist = [] for elm in root.getiterator("%spackage" % ns): if elm.find("%sname" % ns).text == 'rpm': for node in elm.getchildren(): if node.tag == "%sversion" % ns: versionlist.append(node.attrib['ver']) if versionlist: return reversed( sorted( versionlist, key = lambda ver: map(int, ver.split('.')))).next() elif repo["primary"].endswith(".sqlite"): con = sqlite.connect(repo["primary"]) for row in con.execute("select version from packages where " "name=\"rpm\" ORDER by version DESC"): con.close() return row[0] return None def get_arch(repometadata): archlist = [] for repo in repometadata: if repo["primary"].endswith(".xml"): root = xmlparse(repo["primary"]) ns = root.getroot().tag ns = ns[0:ns.rindex("}")+1] for elm in root.getiterator("%spackage" % ns): if elm.find("%sarch" % ns).text not in ("noarch", "src"): arch = elm.find("%sarch" % ns).text if arch not in archlist: archlist.append(arch) elif repo["primary"].endswith(".sqlite"): con = sqlite.connect(repo["primary"]) for row in con.execute("select arch from packages where arch not in (\"src\", \"noarch\")"): if row[0] not in archlist: archlist.append(row[0]) con.close() uniq_arch = [] for i in range(len(archlist)): if archlist[i] not in rpmmisc.archPolicies.keys(): continue need_append = True j = 0 while j < len(uniq_arch): if archlist[i] in rpmmisc.archPolicies[uniq_arch[j]].split(':'): need_append = False break if uniq_arch[j] in rpmmisc.archPolicies[archlist[i]].split(':'): if need_append: uniq_arch[j] = archlist[i] need_append = False else: uniq_arch.remove(uniq_arch[j]) continue j += 1 if need_append: uniq_arch.append(archlist[i]) return uniq_arch, archlist def get_package(pkg, repometadata, arch = None): ver = "" target_repo = None if not arch: arches = [] elif arch not in rpmmisc.archPolicies: arches = [arch] else: arches = rpmmisc.archPolicies[arch].split(':') arches.append('noarch') for repo in repometadata: if repo["primary"].endswith(".xml"): root = xmlparse(repo["primary"]) ns = root.getroot().tag ns = ns[0:ns.rindex("}")+1] for elm in root.getiterator("%spackage" % ns): if elm.find("%sname" % ns).text == pkg: if elm.find("%sarch" % ns).text in arches: version = elm.find("%sversion" % ns) tmpver = "%s-%s" % (version.attrib['ver'], version.attrib['rel']) if tmpver > ver: ver = tmpver location = elm.find("%slocation" % ns) pkgpath = "%s" % location.attrib['href'] target_repo = repo break if repo["primary"].endswith(".sqlite"): con = sqlite.connect(repo["primary"]) if arch: sql = 'select version, release, location_href from packages ' \ 'where name = "%s" and arch IN ("%s")' % \ (pkg, '","'.join(arches)) for row in con.execute(sql): tmpver = "%s-%s" % (row[0], row[1]) if tmpver > ver: ver = tmpver pkgpath = "%s" % row[2] target_repo = repo break else: sql = 'select version, release, location_href from packages ' \ 'where name = "%s"' % pkg for row in con.execute(sql): tmpver = "%s-%s" % (row[0], row[1]) if tmpver > ver: ver = tmpver pkgpath = "%s" % row[2] target_repo = repo break con.close() if target_repo: makedirs("%s/packages/%s" % (target_repo["cachedir"], target_repo["name"])) url = os.path.join(target_repo["baseurl"], pkgpath) filename = str("%s/packages/%s/%s" % (target_repo["cachedir"], target_repo["name"], os.path.basename(pkgpath))) if os.path.exists(filename): ret = rpmmisc.checkRpmIntegrity('rpm', filename) if ret == 0: return filename msger.warning("package %s is damaged: %s" % (os.path.basename(filename), filename)) os.unlink(filename) pkg = myurlgrab(str(url), filename, target_repo["proxies"]) return pkg else: return None def get_source_name(pkg, repometadata): def get_bin_name(pkg): m = RPM_RE.match(pkg) if m: return m.group(1) return None def get_src_name(srpm): m = SRPM_RE.match(srpm) if m: return m.group(1) return None ver = "" target_repo = None pkg_name = get_bin_name(pkg) if not pkg_name: return None for repo in repometadata: if repo["primary"].endswith(".xml"): root = xmlparse(repo["primary"]) ns = root.getroot().tag ns = ns[0:ns.rindex("}")+1] for elm in root.getiterator("%spackage" % ns): if elm.find("%sname" % ns).text == pkg_name: if elm.find("%sarch" % ns).text != "src": version = elm.find("%sversion" % ns) tmpver = "%s-%s" % (version.attrib['ver'], version.attrib['rel']) if tmpver > ver: ver = tmpver fmt = elm.find("%sformat" % ns) if fmt: fns = fmt.getchildren()[0].tag fns = fns[0:fns.rindex("}")+1] pkgpath = fmt.find("%ssourcerpm" % fns).text target_repo = repo break if repo["primary"].endswith(".sqlite"): con = sqlite.connect(repo["primary"]) for row in con.execute("select version, release, rpm_sourcerpm from packages where name = \"%s\" and arch != \"src\"" % pkg_name): tmpver = "%s-%s" % (row[0], row[1]) if tmpver > ver: pkgpath = "%s" % row[2] target_repo = repo break con.close() if target_repo: return get_src_name(pkgpath) else: return None def get_pkglist_in_patterns(group, patterns): found = False pkglist = [] try: root = xmlparse(patterns) except SyntaxError: raise SyntaxError("%s syntax error." % patterns) for elm in list(root.getroot()): ns = elm.tag ns = ns[0:ns.rindex("}")+1] name = elm.find("%sname" % ns) summary = elm.find("%ssummary" % ns) if name.text == group or summary.text == group: found = True break if not found: return pkglist found = False for requires in list(elm): if requires.tag.endswith("requires"): found = True break if not found: return pkglist for pkg in list(requires): pkgname = pkg.attrib["name"] if pkgname not in pkglist: pkglist.append(pkgname) return pkglist def get_pkglist_in_comps(group, comps): found = False pkglist = [] try: root = xmlparse(comps) except SyntaxError: raise SyntaxError("%s syntax error." % comps) for elm in root.getiterator("group"): id = elm.find("id") name = elm.find("name") if id.text == group or name.text == group: packagelist = elm.find("packagelist") found = True break if not found: return pkglist for require in elm.getiterator("packagereq"): if require.tag.endswith("packagereq"): pkgname = require.text if pkgname not in pkglist: pkglist.append(pkgname) return pkglist def is_statically_linked(binary): return ", statically linked, " in runner.outs(['file', binary]) def setup_qemu_emulator(rootdir, arch): # mount binfmt_misc if it doesn't exist if not os.path.exists("/proc/sys/fs/binfmt_misc"): modprobecmd = find_binary_path("modprobe") runner.show([modprobecmd, "binfmt_misc"]) if not os.path.exists("/proc/sys/fs/binfmt_misc/register"): mountcmd = find_binary_path("mount") runner.show([mountcmd, "-t", "binfmt_misc", "none", "/proc/sys/fs/binfmt_misc"]) # qemu_emulator is a special case, we can't use find_binary_path # qemu emulator should be a statically-linked executable file qemu_emulator = "/usr/bin/qemu-arm" if not os.path.exists(qemu_emulator) or not is_statically_linked(qemu_emulator): qemu_emulator = "/usr/bin/qemu-arm-static" if not os.path.exists(qemu_emulator): raise CreatorError("Please install a statically-linked qemu-arm") # qemu emulator version check armv7_list = [arch for arch in rpmmisc.archPolicies.keys() if arch.startswith('armv7')] if arch in armv7_list: # need qemu (>=0.13.0) qemuout = runner.outs([qemu_emulator, "-h"]) m = re.search("version\s*([.\d]+)", qemuout) if m: qemu_version = m.group(1) if qemu_version < "0.13": raise CreatorError("Requires %s version >=0.13 for %s" % (qemu_emulator, arch)) else: msger.warning("Can't get version info of %s, please make sure it's higher than 0.13.0" % qemu_emulator) if not os.path.exists(rootdir + "/usr/bin"): makedirs(rootdir + "/usr/bin") shutil.copy(qemu_emulator, rootdir + "/usr/bin/qemu-arm-static") qemu_emulator = "/usr/bin/qemu-arm-static" # disable selinux, selinux will block qemu emulator to run if os.path.exists("/usr/sbin/setenforce"): msger.info('Try to disable selinux') runner.show(["/usr/sbin/setenforce", "0"]) # unregister it if it has been registered and is a dynamically-linked executable node = "/proc/sys/fs/binfmt_misc/arm" if os.path.exists(node): qemu_unregister_string = "-1\n" fd = open("/proc/sys/fs/binfmt_misc/arm", "w") fd.write(qemu_unregister_string) fd.close() # register qemu emulator for interpreting other arch executable file if not os.path.exists(node): qemu_arm_string = ":arm:M::\\x7fELF\\x01\\x01\\x01\\x00\\x00\\x00\\x00\\x00\\x00\\x00\\x00\\x00\\x02\\x00\\x28\\x00:\\xff\\xff\\xff\\xff\\xff\\xff\\xff\\x00\\xff\\xff\\xff\\xff\\xff\\xff\\xff\\xff\\xfa\\xff\\xff\\xff:%s:\n" % qemu_emulator fd = open("/proc/sys/fs/binfmt_misc/register", "w") fd.write(qemu_arm_string) fd.close() return qemu_emulator def SrcpkgsDownload(pkgs, repometadata, instroot, cachedir): def get_source_repometadata(repometadata): src_repometadata=[] for repo in repometadata: if repo["name"].endswith("-source"): src_repometadata.append(repo) if src_repometadata: return src_repometadata return None def get_src_name(srpm): m = SRPM_RE.match(srpm) if m: return m.group(1) return None src_repometadata = get_source_repometadata(repometadata) if not src_repometadata: msger.warning("No source repo found") return None src_pkgs = [] lpkgs_dict = {} lpkgs_path = [] for repo in src_repometadata: cachepath = "%s/%s/packages/*.src.rpm" %(cachedir, repo["name"]) lpkgs_path += glob.glob(cachepath) for lpkg in lpkgs_path: lpkg_name = get_src_name(os.path.basename(lpkg)) lpkgs_dict[lpkg_name] = lpkg localpkgs = lpkgs_dict.keys() cached_count = 0 destdir = instroot+'/usr/src/SRPMS' if not os.path.exists(destdir): os.makedirs(destdir) srcpkgset = set() for _pkg in pkgs: srcpkg_name = get_source_name(_pkg, repometadata) if not srcpkg_name: continue srcpkgset.add(srcpkg_name) for pkg in list(srcpkgset): if pkg in localpkgs: cached_count += 1 shutil.copy(lpkgs_dict[pkg], destdir) src_pkgs.append(os.path.basename(lpkgs_dict[pkg])) else: src_pkg = get_package(pkg, src_repometadata, 'src') if src_pkg: shutil.copy(src_pkg, destdir) src_pkgs.append(src_pkg) msger.info("%d source packages gotten from cache" % cached_count) return src_pkgs def strip_end(text, suffix): if not text.endswith(suffix): return text return text[:-len(suffix)] ########NEW FILE######## __FILENAME__ = misc # ex:ts=4:sw=4:sts=4:et # -*- tab-width: 4; c-basic-offset: 4; indent-tabs-mode: nil -*- # # Copyright (c) 2013, Intel Corporation. # All rights reserved. # # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License version 2 as # published by the Free Software Foundation. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., # 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA. # # DESCRIPTION # This module provides a place to collect various mic-related utils # for the OpenEmbedded Image Tools. # # AUTHORS # Tom Zanussi <tom.zanussi (at] linux.intel.com> # from mic import msger from mic.utils import runner def exec_cmd(cmd_and_args, as_shell = False, catch = 3): """ Execute command, catching stderr, stdout Need to execute as_shell if the command uses wildcards """ msger.debug("exec_cmd: %s" % cmd_and_args) args = cmd_and_args.split() msger.debug(args) if (as_shell): rc, out = runner.runtool(cmd_and_args, catch) else: rc, out = runner.runtool(args, catch) out = out.strip() msger.debug("exec_cmd: output for %s (rc = %d): %s" % \ (cmd_and_args, rc, out)) if rc != 0: # We don't throw exception when return code is not 0, because # parted always fails to reload part table with loop devices. This # prevents us from distinguishing real errors based on return # code. msger.warning("WARNING: %s returned '%s' instead of 0" % (cmd_and_args, rc)) return (rc, out) def exec_cmd_quiet(cmd_and_args, as_shell = False): """ Execute command, catching nothing in the output Need to execute as_shell if the command uses wildcards """ return exec_cmd(cmd_and_args, as_shell, 0) def exec_native_cmd(cmd_and_args, native_sysroot, catch = 3): """ Execute native command, catching stderr, stdout Need to execute as_shell if the command uses wildcards Always need to execute native commands as_shell """ native_paths = \ "export PATH=%s/sbin:%s/usr/sbin:%s/usr/bin:$PATH" % \ (native_sysroot, native_sysroot, native_sysroot) native_cmd_and_args = "%s;%s" % (native_paths, cmd_and_args) msger.debug("exec_native_cmd: %s" % cmd_and_args) args = cmd_and_args.split() msger.debug(args) rc, out = exec_cmd(native_cmd_and_args, True, catch) if rc == 127: # shell command-not-found msger.error("A native (host) program required to build the image " "was not found (see details above). Please make sure " "it's installed and try again.") return (rc, out) def exec_native_cmd_quiet(cmd_and_args, native_sysroot): """ Execute native command, catching nothing in the output Need to execute as_shell if the command uses wildcards Always need to execute native commands as_shell """ return exec_native_cmd(cmd_and_args, native_sysroot, 0) # kickstart doesn't support variable substution in commands, so this # is our current simplistic scheme for supporting that wks_vars = dict() def get_wks_var(key): return wks_vars[key] def add_wks_var(key, val): wks_vars[key] = val BOOTDD_EXTRA_SPACE = 16384 IMAGE_EXTRA_SPACE = 10240 __bitbake_env_lines = "" def set_bitbake_env_lines(bitbake_env_lines): global __bitbake_env_lines __bitbake_env_lines = bitbake_env_lines def get_bitbake_env_lines(): return __bitbake_env_lines def find_bitbake_env_lines(image_name): """ If image_name is empty, plugins might still be able to use the environment, so set it regardless. """ if image_name: bitbake_env_cmd = "bitbake -e %s" % image_name else: bitbake_env_cmd = "bitbake -e" rc, bitbake_env_lines = exec_cmd(bitbake_env_cmd) if rc != 0: print "Couldn't get '%s' output." % bitbake_env_cmd return None return bitbake_env_lines def find_artifact(bitbake_env_lines, variable): """ Gather the build artifact for the current image (the image_name e.g. core-image-minimal) for the current MACHINE set in local.conf """ retval = "" for line in bitbake_env_lines.split('\n'): if (get_line_val(line, variable)): retval = get_line_val(line, variable) break return retval def get_line_val(line, key): """ Extract the value from the VAR="val" string """ if line.startswith(key + "="): stripped_line = line.split('=')[1] stripped_line = stripped_line.replace('\"', '') return stripped_line return None def get_bitbake_var(key): for line in __bitbake_env_lines.split('\n'): if (get_line_val(line, key)): val = get_line_val(line, key) return val return None ########NEW FILE######## __FILENAME__ = partitionedfs #!/usr/bin/python -tt # # Copyright (c) 2009, 2010, 2011 Intel, Inc. # Copyright (c) 2007, 2008 Red Hat, Inc. # Copyright (c) 2008 Daniel P. Berrange # Copyright (c) 2008 David P. Huff # # This program is free software; you can redistribute it and/or modify it # under the terms of the GNU General Public License as published by the Free # Software Foundation; version 2 of the License # # This program is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY # or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License # for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., 59 # Temple Place - Suite 330, Boston, MA 02111-1307, USA. import os from mic import msger from mic.utils import runner from mic.utils.errors import MountError from mic.utils.fs_related import * from mic.utils.gpt_parser import GptParser from mic.utils.oe.misc import * # Overhead of the MBR partitioning scheme (just one sector) MBR_OVERHEAD = 1 # Overhead of the GPT partitioning scheme GPT_OVERHEAD = 34 # Size of a sector in bytes SECTOR_SIZE = 512 class PartitionedMount(Mount): def __init__(self, mountdir, skipformat = False): Mount.__init__(self, mountdir) self.disks = {} self.partitions = [] self.subvolumes = [] self.mapped = False self.mountOrder = [] self.unmountOrder = [] self.parted = find_binary_path("parted") self.btrfscmd=None self.skipformat = skipformat self.snapshot_created = self.skipformat # Size of a sector used in calculations self.sector_size = SECTOR_SIZE self._partitions_layed_out = False def __add_disk(self, disk_name): """ Add a disk 'disk_name' to the internal list of disks. Note, 'disk_name' is the name of the disk in the target system (e.g., sdb). """ if disk_name in self.disks: # We already have this disk return assert not self._partitions_layed_out self.disks[disk_name] = \ { 'disk': None, # Disk object 'mapped': False, # True if kpartx mapping exists 'numpart': 0, # Number of allocate partitions 'partitions': [], # Indexes to self.partitions 'offset': 0, # Offset of next partition (in sectors) # Minimum required disk size to fit all partitions (in bytes) 'min_size': 0, 'ptable_format': "msdos" } # Partition table format def add_disk(self, disk_name, disk_obj): """ Add a disk object which have to be partitioned. More than one disk can be added. In case of multiple disks, disk partitions have to be added for each disk separately with 'add_partition()". """ self.__add_disk(disk_name) self.disks[disk_name]['disk'] = disk_obj def __add_partition(self, part): """ This is a helper function for 'add_partition()' which adds a partition to the internal list of partitions. """ assert not self._partitions_layed_out self.partitions.append(part) self.__add_disk(part['disk_name']) def add_partition(self, size, disk_name, mountpoint, source_file = None, fstype = None, label=None, fsopts = None, boot = False, align = None, part_type = None): """ Add the next partition. Prtitions have to be added in the first-to-last order. """ ks_pnum = len(self.partitions) # Converting MB to sectors for parted size = size * 1024 * 1024 / self.sector_size # We need to handle subvolumes for btrfs if fstype == "btrfs" and fsopts and fsopts.find("subvol=") != -1: self.btrfscmd=find_binary_path("btrfs") subvol = None opts = fsopts.split(",") for opt in opts: if opt.find("subvol=") != -1: subvol = opt.replace("subvol=", "").strip() break if not subvol: raise MountError("No subvolume: %s" % fsopts) self.subvolumes.append({'size': size, # In sectors 'mountpoint': mountpoint, # Mount relative to chroot 'fstype': fstype, # Filesystem type 'fsopts': fsopts, # Filesystem mount options 'disk_name': disk_name, # physical disk name holding partition 'device': None, # kpartx device node for partition 'mount': None, # Mount object 'subvol': subvol, # Subvolume name 'boot': boot, # Bootable flag 'mounted': False # Mount flag }) # We still need partition for "/" or non-subvolume if mountpoint == "/" or not fsopts or fsopts.find("subvol=") == -1: # Don't need subvolume for "/" because it will be set as default subvolume if fsopts and fsopts.find("subvol=") != -1: opts = fsopts.split(",") for opt in opts: if opt.strip().startswith("subvol="): opts.remove(opt) break fsopts = ",".join(opts) part = { 'ks_pnum' : ks_pnum, # Partition number in the KS file 'size': size, # In sectors 'mountpoint': mountpoint, # Mount relative to chroot 'source_file': source_file, # partition contents 'fstype': fstype, # Filesystem type 'fsopts': fsopts, # Filesystem mount options 'label': label, # Partition label 'disk_name': disk_name, # physical disk name holding partition 'device': None, # kpartx device node for partition 'mount': None, # Mount object 'num': None, # Partition number 'boot': boot, # Bootable flag 'align': align, # Partition alignment 'part_type' : part_type, # Partition type 'partuuid': None } # Partition UUID (GPT-only) self.__add_partition(part) def layout_partitions(self, ptable_format = "msdos"): """ Layout the partitions, meaning calculate the position of every partition on the disk. The 'ptable_format' parameter defines the partition table format, and may be either "msdos" or "gpt". """ msger.debug("Assigning %s partitions to disks" % ptable_format) if ptable_format not in ('msdos', 'gpt'): raise MountError("Unknown partition table format '%s', supported " \ "formats are: 'msdos' and 'gpt'" % ptable_format) if self._partitions_layed_out: return self._partitions_layed_out = True # Go through partitions in the order they are added in .ks file for n in range(len(self.partitions)): p = self.partitions[n] if not self.disks.has_key(p['disk_name']): raise MountError("No disk %s for partition %s" \ % (p['disk_name'], p['mountpoint'])) if p['part_type'] and ptable_format != 'gpt': # The --part-type can also be implemented for MBR partitions, # in which case it would map to the 1-byte "partition type" # filed at offset 3 of the partition entry. raise MountError("setting custom partition type is only " \ "imlemented for GPT partitions") # Get the disk where the partition is located d = self.disks[p['disk_name']] d['numpart'] += 1 d['ptable_format'] = ptable_format if d['numpart'] == 1: if ptable_format == "msdos": overhead = MBR_OVERHEAD else: overhead = GPT_OVERHEAD # Skip one sector required for the partitioning scheme overhead d['offset'] += overhead # Steal few sectors from the first partition to offset for the # partitioning overhead p['size'] -= overhead if p['align']: # If not first partition and we do have alignment set we need # to align the partition. # FIXME: This leaves a empty spaces to the disk. To fill the # gaps we could enlargea the previous partition? # Calc how much the alignment is off. align_sectors = d['offset'] % (p['align'] * 1024 / self.sector_size) # We need to move forward to the next alignment point align_sectors = (p['align'] * 1024 / self.sector_size) - align_sectors msger.debug("Realignment for %s%s with %s sectors, original" " offset %s, target alignment is %sK." % (p['disk_name'], d['numpart'], align_sectors, d['offset'], p['align'])) # increase the offset so we actually start the partition on right alignment d['offset'] += align_sectors p['start'] = d['offset'] d['offset'] += p['size'] p['type'] = 'primary' p['num'] = d['numpart'] if d['ptable_format'] == "msdos": if d['numpart'] > 2: # Every logical partition requires an additional sector for # the EBR, so steal the last sector from the end of each # partition starting from the 3rd one for the EBR. This # will make sure the logical partitions are aligned # correctly. p['size'] -= 1 if d['numpart'] > 3: p['type'] = 'logical' p['num'] = d['numpart'] + 1 d['partitions'].append(n) msger.debug("Assigned %s to %s%d, sectors range %d-%d size %d " "sectors (%d bytes)." \ % (p['mountpoint'], p['disk_name'], p['num'], p['start'], p['start'] + p['size'] - 1, p['size'], p['size'] * self.sector_size)) # Once all the partitions have been layed out, we can calculate the # minumim disk sizes. for disk_name, d in self.disks.items(): d['min_size'] = d['offset'] if d['ptable_format'] == 'gpt': # Account for the backup partition table at the end of the disk d['min_size'] += GPT_OVERHEAD d['min_size'] *= self.sector_size def __run_parted(self, args): """ Run parted with arguments specified in the 'args' list. """ args.insert(0, self.parted) msger.debug(args) rc, out = runner.runtool(args, catch = 3) out = out.strip() if out: msger.debug('"parted" output: %s' % out) if rc != 0: # We don't throw exception when return code is not 0, because # parted always fails to reload part table with loop devices. This # prevents us from distinguishing real errors based on return # code. msger.debug("WARNING: parted returned '%s' instead of 0" % rc) def __create_partition(self, device, parttype, fstype, start, size): """ Create a partition on an image described by the 'device' object. """ # Start is included to the size so we need to substract one from the end. end = start + size - 1 msger.debug("Added '%s' partition, sectors %d-%d, size %d sectors" % (parttype, start, end, size)) args = ["-s", device, "unit", "s", "mkpart", parttype] if fstype: args.extend([fstype]) args.extend(["%d" % start, "%d" % end]) return self.__run_parted(args) def __format_disks(self): self.layout_partitions() if self.skipformat: msger.debug("Skipping disk format, because skipformat flag is set.") return for dev in self.disks.keys(): d = self.disks[dev] msger.debug("Initializing partition table for %s" % \ (d['disk'].device)) self.__run_parted(["-s", d['disk'].device, "mklabel", d['ptable_format']]) msger.debug("Creating partitions") for p in self.partitions: d = self.disks[p['disk_name']] if d['ptable_format'] == "msdos" and p['num'] == 5: # The last sector of the 3rd partition was reserved for the EBR # of the first _logical_ partition. This is why the extended # partition should start one sector before the first logical # partition. self.__create_partition(d['disk'].device, "extended", None, p['start'] - 1, d['offset'] - p['start']) if p['fstype'] == "swap": parted_fs_type = "linux-swap" elif p['fstype'] == "vfat": parted_fs_type = "fat32" elif p['fstype'] == "msdos": parted_fs_type = "fat16" else: # Type for ext2/ext3/ext4/btrfs parted_fs_type = "ext2" # Boot ROM of OMAP boards require vfat boot partition to have an # even number of sectors. if p['mountpoint'] == "/boot" and p['fstype'] in ["vfat", "msdos"] \ and p['size'] % 2: msger.debug("Substracting one sector from '%s' partition to " \ "get even number of sectors for the partition" % \ p['mountpoint']) p['size'] -= 1 self.__create_partition(d['disk'].device, p['type'], parted_fs_type, p['start'], p['size']) if p['boot']: if d['ptable_format'] == 'gpt': flag_name = "legacy_boot" else: flag_name = "boot" msger.debug("Set '%s' flag for partition '%s' on disk '%s'" % \ (flag_name, p['num'], d['disk'].device)) self.__run_parted(["-s", d['disk'].device, "set", "%d" % p['num'], flag_name, "on"]) # Parted defaults to enabling the lba flag for fat16 partitions, # which causes compatibility issues with some firmware (and really # isn't necessary). if parted_fs_type == "fat16": if d['ptable_format'] == 'msdos': msger.debug("Disable 'lba' flag for partition '%s' on disk '%s'" % \ (p['num'], d['disk'].device)) self.__run_parted(["-s", d['disk'].device, "set", "%d" % p['num'], "lba", "off"]) # If the partition table format is "gpt", find out PARTUUIDs for all # the partitions. And if users specified custom parition type UUIDs, # set them. for disk_name, disk in self.disks.items(): if disk['ptable_format'] != 'gpt': continue pnum = 0 gpt_parser = GptParser(d['disk'].device, SECTOR_SIZE) # Iterate over all GPT partitions on this disk for entry in gpt_parser.get_partitions(): pnum += 1 # Find the matching partition in the 'self.partitions' list for n in d['partitions']: p = self.partitions[n] if p['num'] == pnum: # Found, fetch PARTUUID (partition's unique ID) p['partuuid'] = entry['part_uuid'] msger.debug("PARTUUID for partition %d on disk '%s' " \ "(mount point '%s') is '%s'" % (pnum, \ disk_name, p['mountpoint'], p['partuuid'])) if p['part_type']: entry['type_uuid'] = p['part_type'] msger.debug("Change type of partition %d on disk " \ "'%s' (mount point '%s') to '%s'" % \ (pnum, disk_name, p['mountpoint'], p['part_type'])) gpt_parser.change_partition(entry) del gpt_parser def __map_partitions(self): """Load it if dm_snapshot isn't loaded. """ load_module("dm_snapshot") for dev in self.disks.keys(): d = self.disks[dev] if d['mapped']: continue msger.debug("Running kpartx on %s" % d['disk'].device ) rc, kpartxOutput = runner.runtool([self.kpartx, "-l", "-v", d['disk'].device]) kpartxOutput = kpartxOutput.splitlines() if rc != 0: raise MountError("Failed to query partition mapping for '%s'" % d['disk'].device) # Strip trailing blank and mask verbose output i = 0 while i < len(kpartxOutput) and kpartxOutput[i][0:4] != "loop": i = i + 1 kpartxOutput = kpartxOutput[i:] # Make sure kpartx reported the right count of partitions if len(kpartxOutput) != d['numpart']: # If this disk has more than 3 partitions, then in case of MBR # paritions there is an extended parition. Different versions # of kpartx behave differently WRT the extended partition - # some map it, some ignore it. This is why we do the below hack # - if kpartx reported one more partition and the partition # table type is "msdos" and the amount of partitions is more # than 3, we just assume kpartx mapped the extended parition # and we remove it. if len(kpartxOutput) == d['numpart'] + 1 \ and d['ptable_format'] == 'msdos' and len(kpartxOutput) > 3: kpartxOutput.pop(3) else: raise MountError("Unexpected number of partitions from " \ "kpartx: %d != %d" % \ (len(kpartxOutput), d['numpart'])) for i in range(len(kpartxOutput)): line = kpartxOutput[i] newdev = line.split()[0] mapperdev = "/dev/mapper/" + newdev loopdev = d['disk'].device + newdev[-1] msger.debug("Dev %s: %s -> %s" % (newdev, loopdev, mapperdev)) pnum = d['partitions'][i] self.partitions[pnum]['device'] = loopdev # grub's install wants partitions to be named # to match their parent device + partition num # kpartx doesn't work like this, so we add compat # symlinks to point to /dev/mapper if os.path.lexists(loopdev): os.unlink(loopdev) os.symlink(mapperdev, loopdev) msger.debug("Adding partx mapping for %s" % d['disk'].device) rc = runner.show([self.kpartx, "-v", "-a", d['disk'].device]) if rc != 0: # Make sure that the device maps are also removed on error case. # The d['mapped'] isn't set to True if the kpartx fails so # failed mapping will not be cleaned on cleanup either. runner.quiet([self.kpartx, "-d", d['disk'].device]) raise MountError("Failed to map partitions for '%s'" % d['disk'].device) # FIXME: there is a bit delay for multipath device setup, # wait 10ms for the setup import time time.sleep(10) d['mapped'] = True def __unmap_partitions(self): for dev in self.disks.keys(): d = self.disks[dev] if not d['mapped']: continue msger.debug("Removing compat symlinks") for pnum in d['partitions']: if self.partitions[pnum]['device'] != None: os.unlink(self.partitions[pnum]['device']) self.partitions[pnum]['device'] = None msger.debug("Unmapping %s" % d['disk'].device) rc = runner.quiet([self.kpartx, "-d", d['disk'].device]) if rc != 0: raise MountError("Failed to unmap partitions for '%s'" % d['disk'].device) d['mapped'] = False def __calculate_mountorder(self): msger.debug("Calculating mount order") for p in self.partitions: if p['mountpoint']: self.mountOrder.append(p['mountpoint']) self.unmountOrder.append(p['mountpoint']) self.mountOrder.sort() self.unmountOrder.sort() self.unmountOrder.reverse() def cleanup(self): Mount.cleanup(self) if self.disks: self.__unmap_partitions() for dev in self.disks.keys(): d = self.disks[dev] try: d['disk'].cleanup() except: pass def unmount(self): self.__unmount_subvolumes() for mp in self.unmountOrder: if mp == 'swap': continue p = None for p1 in self.partitions: if p1['mountpoint'] == mp: p = p1 break if p['mount'] != None: try: # Create subvolume snapshot here if p['fstype'] == "btrfs" and p['mountpoint'] == "/" and not self.snapshot_created: self.__create_subvolume_snapshots(p, p["mount"]) p['mount'].cleanup() except: pass p['mount'] = None # Only for btrfs def __get_subvolume_id(self, rootpath, subvol): if not self.btrfscmd: self.btrfscmd=find_binary_path("btrfs") argv = [ self.btrfscmd, "subvolume", "list", rootpath ] rc, out = runner.runtool(argv) msger.debug(out) if rc != 0: raise MountError("Failed to get subvolume id from %s', return code: %d." % (rootpath, rc)) subvolid = -1 for line in out.splitlines(): if line.endswith(" path %s" % subvol): subvolid = line.split()[1] if not subvolid.isdigit(): raise MountError("Invalid subvolume id: %s" % subvolid) subvolid = int(subvolid) break return subvolid def __create_subvolume_metadata(self, p, pdisk): if len(self.subvolumes) == 0: return argv = [ self.btrfscmd, "subvolume", "list", pdisk.mountdir ] rc, out = runner.runtool(argv) msger.debug(out) if rc != 0: raise MountError("Failed to get subvolume id from %s', return code: %d." % (pdisk.mountdir, rc)) subvolid_items = out.splitlines() subvolume_metadata = "" for subvol in self.subvolumes: for line in subvolid_items: if line.endswith(" path %s" % subvol["subvol"]): subvolid = line.split()[1] if not subvolid.isdigit(): raise MountError("Invalid subvolume id: %s" % subvolid) subvolid = int(subvolid) opts = subvol["fsopts"].split(",") for opt in opts: if opt.strip().startswith("subvol="): opts.remove(opt) break fsopts = ",".join(opts) subvolume_metadata += "%d\t%s\t%s\t%s\n" % (subvolid, subvol["subvol"], subvol['mountpoint'], fsopts) if subvolume_metadata: fd = open("%s/.subvolume_metadata" % pdisk.mountdir, "w") fd.write(subvolume_metadata) fd.close() def __get_subvolume_metadata(self, p, pdisk): subvolume_metadata_file = "%s/.subvolume_metadata" % pdisk.mountdir if not os.path.exists(subvolume_metadata_file): return fd = open(subvolume_metadata_file, "r") content = fd.read() fd.close() for line in content.splitlines(): items = line.split("\t") if items and len(items) == 4: self.subvolumes.append({'size': 0, # In sectors 'mountpoint': items[2], # Mount relative to chroot 'fstype': "btrfs", # Filesystem type 'fsopts': items[3] + ",subvol=%s" % items[1], # Filesystem mount options 'disk_name': p['disk_name'], # physical disk name holding partition 'device': None, # kpartx device node for partition 'mount': None, # Mount object 'subvol': items[1], # Subvolume name 'boot': False, # Bootable flag 'mounted': False # Mount flag }) def __create_subvolumes(self, p, pdisk): """ Create all the subvolumes. """ for subvol in self.subvolumes: argv = [ self.btrfscmd, "subvolume", "create", pdisk.mountdir + "/" + subvol["subvol"]] rc = runner.show(argv) if rc != 0: raise MountError("Failed to create subvolume '%s', return code: %d." % (subvol["subvol"], rc)) # Set default subvolume, subvolume for "/" is default subvol = None for subvolume in self.subvolumes: if subvolume["mountpoint"] == "/" and p['disk_name'] == subvolume['disk_name']: subvol = subvolume break if subvol: # Get default subvolume id subvolid = self. __get_subvolume_id(pdisk.mountdir, subvol["subvol"]) # Set default subvolume if subvolid != -1: rc = runner.show([ self.btrfscmd, "subvolume", "set-default", "%d" % subvolid, pdisk.mountdir]) if rc != 0: raise MountError("Failed to set default subvolume id: %d', return code: %d." % (subvolid, rc)) self.__create_subvolume_metadata(p, pdisk) def __mount_subvolumes(self, p, pdisk): if self.skipformat: # Get subvolume info self.__get_subvolume_metadata(p, pdisk) # Set default mount options if len(self.subvolumes) != 0: for subvol in self.subvolumes: if subvol["mountpoint"] == p["mountpoint"] == "/": opts = subvol["fsopts"].split(",") for opt in opts: if opt.strip().startswith("subvol="): opts.remove(opt) break pdisk.fsopts = ",".join(opts) break if len(self.subvolumes) == 0: # Return directly if no subvolumes return # Remount to make default subvolume mounted rc = runner.show([self.umountcmd, pdisk.mountdir]) if rc != 0: raise MountError("Failed to umount %s" % pdisk.mountdir) rc = runner.show([self.mountcmd, "-o", pdisk.fsopts, pdisk.disk.device, pdisk.mountdir]) if rc != 0: raise MountError("Failed to umount %s" % pdisk.mountdir) for subvol in self.subvolumes: if subvol["mountpoint"] == "/": continue subvolid = self. __get_subvolume_id(pdisk.mountdir, subvol["subvol"]) if subvolid == -1: msger.debug("WARNING: invalid subvolume %s" % subvol["subvol"]) continue # Replace subvolume name with subvolume ID opts = subvol["fsopts"].split(",") for opt in opts: if opt.strip().startswith("subvol="): opts.remove(opt) break opts.extend(["subvolrootid=0", "subvol=%s" % subvol["subvol"]]) fsopts = ",".join(opts) subvol['fsopts'] = fsopts mountpoint = self.mountdir + subvol['mountpoint'] makedirs(mountpoint) rc = runner.show([self.mountcmd, "-o", fsopts, pdisk.disk.device, mountpoint]) if rc != 0: raise MountError("Failed to mount subvolume %s to %s" % (subvol["subvol"], mountpoint)) subvol["mounted"] = True def __unmount_subvolumes(self): """ It may be called multiple times, so we need to chekc if it is still mounted. """ for subvol in self.subvolumes: if subvol["mountpoint"] == "/": continue if not subvol["mounted"]: continue mountpoint = self.mountdir + subvol['mountpoint'] rc = runner.show([self.umountcmd, mountpoint]) if rc != 0: raise MountError("Failed to unmount subvolume %s from %s" % (subvol["subvol"], mountpoint)) subvol["mounted"] = False def __create_subvolume_snapshots(self, p, pdisk): import time if self.snapshot_created: return # Remount with subvolid=0 rc = runner.show([self.umountcmd, pdisk.mountdir]) if rc != 0: raise MountError("Failed to umount %s" % pdisk.mountdir) if pdisk.fsopts: mountopts = pdisk.fsopts + ",subvolid=0" else: mountopts = "subvolid=0" rc = runner.show([self.mountcmd, "-o", mountopts, pdisk.disk.device, pdisk.mountdir]) if rc != 0: raise MountError("Failed to umount %s" % pdisk.mountdir) # Create all the subvolume snapshots snapshotts = time.strftime("%Y%m%d-%H%M") for subvol in self.subvolumes: subvolpath = pdisk.mountdir + "/" + subvol["subvol"] snapshotpath = subvolpath + "_%s-1" % snapshotts rc = runner.show([ self.btrfscmd, "subvolume", "snapshot", subvolpath, snapshotpath ]) if rc != 0: raise MountError("Failed to create subvolume snapshot '%s' for '%s', return code: %d." % (snapshotpath, subvolpath, rc)) self.snapshot_created = True def __install_partition(self, num, source_file, start, size): """ Install source_file contents into a partition. """ if not source_file: # nothing to install return # Start is included in the size so need to substract one from the end. end = start + size - 1 msger.debug("Installed %s in partition %d, sectors %d-%d, size %d sectors" % (source_file, num, start, end, size)) dd_cmd = "dd if=%s of=%s bs=%d seek=%d count=%d conv=notrunc" % \ (source_file, self.image_file, self.sector_size, start, size) rc, out = exec_cmd(dd_cmd) def install(self, image_file): msger.debug("Installing partitions") self.image_file = image_file for p in self.partitions: d = self.disks[p['disk_name']] if d['ptable_format'] == "msdos" and p['num'] == 5: # The last sector of the 3rd partition was reserved for the EBR # of the first _logical_ partition. This is why the extended # partition should start one sector before the first logical # partition. self.__install_partition(p['num'], p['source_file'], p['start'] - 1, d['offset'] - p['start']) self.__install_partition(p['num'], p['source_file'], p['start'], p['size']) def mount(self): for dev in self.disks.keys(): d = self.disks[dev] d['disk'].create() self.__format_disks() self.__calculate_mountorder() return def resparse(self, size = None): # Can't re-sparse a disk image - too hard pass ########NEW FILE######## __FILENAME__ = proxy #!/usr/bin/python -tt # # Copyright (c) 2010, 2011 Intel, Inc. # # This program is free software; you can redistribute it and/or modify it # under the terms of the GNU General Public License as published by the Free # Software Foundation; version 2 of the License # # This program is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY # or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License # for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., 59 # Temple Place - Suite 330, Boston, MA 02111-1307, USA. import os import urlparse _my_proxies = {} _my_noproxy = None _my_noproxy_list = [] def set_proxy_environ(): global _my_noproxy, _my_proxies if not _my_proxies: return for key in _my_proxies.keys(): os.environ[key + "_proxy"] = _my_proxies[key] if not _my_noproxy: return os.environ["no_proxy"] = _my_noproxy def unset_proxy_environ(): for env in ('http_proxy', 'https_proxy', 'ftp_proxy', 'all_proxy'): if env in os.environ: del os.environ[env] ENV=env.upper() if ENV in os.environ: del os.environ[ENV] def _set_proxies(proxy = None, no_proxy = None): """Return a dictionary of scheme -> proxy server URL mappings. """ global _my_noproxy, _my_proxies _my_proxies = {} _my_noproxy = None proxies = [] if proxy: proxies.append(("http_proxy", proxy)) if no_proxy: proxies.append(("no_proxy", no_proxy)) # Get proxy settings from environment if not provided if not proxy and not no_proxy: proxies = os.environ.items() # Remove proxy env variables, urllib2 can't handle them correctly unset_proxy_environ() for name, value in proxies: name = name.lower() if value and name[-6:] == '_proxy': if name[0:2] != "no": _my_proxies[name[:-6]] = value else: _my_noproxy = value def _ip_to_int(ip): ipint=0 shift=24 for dec in ip.split("."): ipint |= int(dec) << shift shift -= 8 return ipint def _int_to_ip(val): ipaddr="" shift=0 for i in range(4): dec = val >> shift dec &= 0xff ipaddr = ".%d%s" % (dec, ipaddr) shift += 8 return ipaddr[1:] def _isip(host): if host.replace(".", "").isdigit(): return True return False def _set_noproxy_list(): global _my_noproxy, _my_noproxy_list _my_noproxy_list = [] if not _my_noproxy: return for item in _my_noproxy.split(","): item = item.strip() if not item: continue if item[0] != '.' and item.find("/") == -1: # Need to match it _my_noproxy_list.append({"match":0,"needle":item}) elif item[0] == '.': # Need to match at tail _my_noproxy_list.append({"match":1,"needle":item}) elif item.find("/") > 3: # IP/MASK, need to match at head needle = item[0:item.find("/")].strip() ip = _ip_to_int(needle) netmask = 0 mask = item[item.find("/")+1:].strip() if mask.isdigit(): netmask = int(mask) netmask = ~((1<<(32-netmask)) - 1) ip &= netmask else: shift=24 netmask=0 for dec in mask.split("."): netmask |= int(dec) << shift shift -= 8 ip &= netmask _my_noproxy_list.append({"match":2,"needle":ip,"netmask":netmask}) def _isnoproxy(url): (scheme, host, path, parm, query, frag) = urlparse.urlparse(url) if '@' in host: user_pass, host = host.split('@', 1) if ':' in host: host, port = host.split(':', 1) hostisip = _isip(host) for item in _my_noproxy_list: if hostisip and item["match"] <= 1: continue if item["match"] == 2 and hostisip: if (_ip_to_int(host) & item["netmask"]) == item["needle"]: return True if item["match"] == 0: if host == item["needle"]: return True if item["match"] == 1: if host.rfind(item["needle"]) > 0: return True return False def set_proxies(proxy = None, no_proxy = None): _set_proxies(proxy, no_proxy) _set_noproxy_list() set_proxy_environ() def get_proxy_for(url): if url.startswith('file:') or _isnoproxy(url): return None type = url[0:url.index(":")] proxy = None if _my_proxies.has_key(type): proxy = _my_proxies[type] elif _my_proxies.has_key("http"): proxy = _my_proxies["http"] else: proxy = None return proxy ########NEW FILE######## __FILENAME__ = rpmmisc #!/usr/bin/python -tt # # Copyright (c) 2008, 2009, 2010, 2011 Intel, Inc. # # This program is free software; you can redistribute it and/or modify it # under the terms of the GNU General Public License as published by the Free # Software Foundation; version 2 of the License # # This program is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY # or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License # for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., 59 # Temple Place - Suite 330, Boston, MA 02111-1307, USA. import os import sys import re import rpm from mic import msger from mic.utils.errors import CreatorError from mic.utils.proxy import get_proxy_for from mic.utils import runner class RPMInstallCallback: """ Command line callback class for callbacks from the RPM library. """ def __init__(self, ts, output=1): self.output = output self.callbackfilehandles = {} self.total_actions = 0 self.total_installed = 0 self.installed_pkg_names = [] self.total_removed = 0 self.mark = "+" self.marks = 40 self.lastmsg = None self.tsInfo = None # this needs to be set for anything else to work self.ts = ts self.filelog = False self.logString = [] self.headmsg = "Installing" def _dopkgtup(self, hdr): tmpepoch = hdr['epoch'] if tmpepoch is None: epoch = '0' else: epoch = str(tmpepoch) return (hdr['name'], hdr['arch'], epoch, hdr['version'], hdr['release']) def _makeHandle(self, hdr): handle = '%s:%s.%s-%s-%s' % (hdr['epoch'], hdr['name'], hdr['version'], hdr['release'], hdr['arch']) return handle def _localprint(self, msg): if self.output: msger.info(msg) def _makefmt(self, percent, progress = True): l = len(str(self.total_actions)) size = "%s.%s" % (l, l) fmt_done = "[%" + size + "s/%" + size + "s]" done = fmt_done % (self.total_installed + self.total_removed, self.total_actions) marks = self.marks - (2 * l) width = "%s.%s" % (marks, marks) fmt_bar = "%-" + width + "s" if progress: bar = fmt_bar % (self.mark * int(marks * (percent / 100.0)), ) fmt = "\r %-10.10s: %-20.20s " + bar + " " + done else: bar = fmt_bar % (self.mark * marks, ) fmt = " %-10.10s: %-20.20s " + bar + " " + done return fmt def _logPkgString(self, hdr): """return nice representation of the package for the log""" (n,a,e,v,r) = self._dopkgtup(hdr) if e == '0': pkg = '%s.%s %s-%s' % (n, a, v, r) else: pkg = '%s.%s %s:%s-%s' % (n, a, e, v, r) return pkg def callback(self, what, bytes, total, h, user): if what == rpm.RPMCALLBACK_TRANS_START: if bytes == 6: self.total_actions = total elif what == rpm.RPMCALLBACK_TRANS_PROGRESS: pass elif what == rpm.RPMCALLBACK_TRANS_STOP: pass elif what == rpm.RPMCALLBACK_INST_OPEN_FILE: self.lastmsg = None hdr = None if h is not None: try: hdr, rpmloc = h except: rpmloc = h hdr = readRpmHeader(self.ts, h) handle = self._makeHandle(hdr) fd = os.open(rpmloc, os.O_RDONLY) self.callbackfilehandles[handle]=fd if hdr['name'] not in self.installed_pkg_names: self.installed_pkg_names.append(hdr['name']) self.total_installed += 1 return fd else: self._localprint("No header - huh?") elif what == rpm.RPMCALLBACK_INST_CLOSE_FILE: hdr = None if h is not None: try: hdr, rpmloc = h except: rpmloc = h hdr = readRpmHeader(self.ts, h) handle = self._makeHandle(hdr) os.close(self.callbackfilehandles[handle]) fd = 0 # log stuff #pkgtup = self._dopkgtup(hdr) self.logString.append(self._logPkgString(hdr)) elif what == rpm.RPMCALLBACK_INST_PROGRESS: if h is not None: percent = (self.total_installed*100L)/self.total_actions if total > 0: try: hdr, rpmloc = h except: rpmloc = h m = re.match("(.*)-(\d+.*)-(\d+\.\d+)\.(.+)\.rpm", os.path.basename(rpmloc)) if m: pkgname = m.group(1) else: pkgname = os.path.basename(rpmloc) if self.output: fmt = self._makefmt(percent) msg = fmt % (self.headmsg, pkgname) if msg != self.lastmsg: self.lastmsg = msg msger.info(msg) if self.total_installed == self.total_actions: msger.raw('') msger.verbose('\n'.join(self.logString)) elif what == rpm.RPMCALLBACK_UNINST_START: pass elif what == rpm.RPMCALLBACK_UNINST_PROGRESS: pass elif what == rpm.RPMCALLBACK_UNINST_STOP: self.total_removed += 1 elif what == rpm.RPMCALLBACK_REPACKAGE_START: pass elif what == rpm.RPMCALLBACK_REPACKAGE_STOP: pass elif what == rpm.RPMCALLBACK_REPACKAGE_PROGRESS: pass def readRpmHeader(ts, filename): """ Read an rpm header. """ fd = os.open(filename, os.O_RDONLY) h = ts.hdrFromFdno(fd) os.close(fd) return h def splitFilename(filename): """ Pass in a standard style rpm fullname Return a name, version, release, epoch, arch, e.g.:: foo-1.0-1.i386.rpm returns foo, 1.0, 1, i386 1:bar-9-123a.ia64.rpm returns bar, 9, 123a, 1, ia64 """ if filename[-4:] == '.rpm': filename = filename[:-4] archIndex = filename.rfind('.') arch = filename[archIndex+1:] relIndex = filename[:archIndex].rfind('-') rel = filename[relIndex+1:archIndex] verIndex = filename[:relIndex].rfind('-') ver = filename[verIndex+1:relIndex] epochIndex = filename.find(':') if epochIndex == -1: epoch = '' else: epoch = filename[:epochIndex] name = filename[epochIndex + 1:verIndex] return name, ver, rel, epoch, arch def getCanonX86Arch(arch): # if arch == "i586": f = open("/proc/cpuinfo", "r") lines = f.readlines() f.close() for line in lines: if line.startswith("model name") and line.find("Geode(TM)") != -1: return "geode" return arch # only athlon vs i686 isn't handled with uname currently if arch != "i686": return arch # if we're i686 and AuthenticAMD, then we should be an athlon f = open("/proc/cpuinfo", "r") lines = f.readlines() f.close() for line in lines: if line.startswith("vendor") and line.find("AuthenticAMD") != -1: return "athlon" # i686 doesn't guarantee cmov, but we depend on it elif line.startswith("flags") and line.find("cmov") == -1: return "i586" return arch def getCanonX86_64Arch(arch): if arch != "x86_64": return arch vendor = None f = open("/proc/cpuinfo", "r") lines = f.readlines() f.close() for line in lines: if line.startswith("vendor_id"): vendor = line.split(':')[1] break if vendor is None: return arch if vendor.find("Authentic AMD") != -1 or vendor.find("AuthenticAMD") != -1: return "amd64" if vendor.find("GenuineIntel") != -1: return "ia32e" return arch def getCanonArch(): arch = os.uname()[4] if (len(arch) == 4 and arch[0] == "i" and arch[2:4] == "86"): return getCanonX86Arch(arch) if arch == "x86_64": return getCanonX86_64Arch(arch) return arch # Copy from libsatsolver:poolarch.c, with cleanup archPolicies = { "x86_64": "x86_64:i686:i586:i486:i386", "i686": "i686:i586:i486:i386", "i586": "i586:i486:i386", "ia64": "ia64:i686:i586:i486:i386", "armv7tnhl": "armv7tnhl:armv7thl:armv7nhl:armv7hl", "armv7thl": "armv7thl:armv7hl", "armv7nhl": "armv7nhl:armv7hl", "armv7hl": "armv7hl", "armv7l": "armv7l:armv6l:armv5tejl:armv5tel:armv5l:armv4tl:armv4l:armv3l", "armv6l": "armv6l:armv5tejl:armv5tel:armv5l:armv4tl:armv4l:armv3l", "armv5tejl": "armv5tejl:armv5tel:armv5l:armv4tl:armv4l:armv3l", "armv5tel": "armv5tel:armv5l:armv4tl:armv4l:armv3l", "armv5l": "armv5l:armv4tl:armv4l:armv3l", } # dict mapping arch -> ( multicompat, best personality, biarch personality ) multilibArches = { "x86_64": ( "athlon", "x86_64", "athlon" ), } # from yumUtils.py arches = { # ia32 "athlon": "i686", "i686": "i586", "geode": "i586", "i586": "i486", "i486": "i386", "i386": "noarch", # amd64 "x86_64": "athlon", "amd64": "x86_64", "ia32e": "x86_64", # arm "armv7tnhl": "armv7nhl", "armv7nhl": "armv7hl", "armv7hl": "noarch", "armv7l": "armv6l", "armv6l": "armv5tejl", "armv5tejl": "armv5tel", "armv5tel": "noarch", #itanium "ia64": "noarch", } def isMultiLibArch(arch=None): """returns true if arch is a multilib arch, false if not""" if arch is None: arch = getCanonArch() if not arches.has_key(arch): # or we could check if it is noarch return False if multilibArches.has_key(arch): return True if multilibArches.has_key(arches[arch]): return True return False def getBaseArch(): myarch = getCanonArch() if not arches.has_key(myarch): return myarch if isMultiLibArch(arch=myarch): if multilibArches.has_key(myarch): return myarch else: return arches[myarch] if arches.has_key(myarch): basearch = myarch value = arches[basearch] while value != 'noarch': basearch = value value = arches[basearch] return basearch def checkRpmIntegrity(bin_rpm, package): return runner.quiet([bin_rpm, "-K", "--nosignature", package]) def checkSig(ts, package): """ Takes a transaction set and a package, check it's sigs, return 0 if they are all fine return 1 if the gpg key can't be found return 2 if the header is in someway damaged return 3 if the key is not trusted return 4 if the pkg is not gpg or pgp signed """ value = 0 currentflags = ts.setVSFlags(0) fdno = os.open(package, os.O_RDONLY) try: hdr = ts.hdrFromFdno(fdno) except rpm.error, e: if str(e) == "public key not availaiable": value = 1 if str(e) == "public key not available": value = 1 if str(e) == "public key not trusted": value = 3 if str(e) == "error reading package header": value = 2 else: error, siginfo = getSigInfo(hdr) if error == 101: os.close(fdno) del hdr value = 4 else: del hdr try: os.close(fdno) except OSError: pass ts.setVSFlags(currentflags) # put things back like they were before return value def getSigInfo(hdr): """ checks signature from an hdr hand back signature information and/or an error code """ import locale locale.setlocale(locale.LC_ALL, 'C') string = '%|DSAHEADER?{%{DSAHEADER:pgpsig}}:{%|RSAHEADER?{%{RSAHEADER:pgpsig}}:{%|SIGGPG?{%{SIGGPG:pgpsig}}:{%|SIGPGP?{%{SIGPGP:pgpsig}}:{(none)}|}|}|}|' siginfo = hdr.sprintf(string) if siginfo != '(none)': error = 0 sigtype, sigdate, sigid = siginfo.split(',') else: error = 101 sigtype = 'MD5' sigdate = 'None' sigid = 'None' infotuple = (sigtype, sigdate, sigid) return error, infotuple def checkRepositoryEULA(name, repo): """ This function is to check the EULA file if provided. return True: no EULA or accepted return False: user declined the EULA """ import tempfile import shutil import urlparse import urllib2 as u2 import httplib from mic.utils.errors import CreatorError def _check_and_download_url(u2opener, url, savepath): try: if u2opener: f = u2opener.open(url) else: f = u2.urlopen(url) except u2.HTTPError, httperror: if httperror.code in (404, 503): return None else: raise CreatorError(httperror) except OSError, oserr: if oserr.errno == 2: return None else: raise CreatorError(oserr) except IOError, oserr: if hasattr(oserr, "reason") and oserr.reason.errno == 2: return None else: raise CreatorError(oserr) except u2.URLError, err: raise CreatorError(err) except httplib.HTTPException, e: raise CreatorError(e) # save to file licf = open(savepath, "w") licf.write(f.read()) licf.close() f.close() return savepath def _pager_file(savepath): if os.path.splitext(savepath)[1].upper() in ('.HTM', '.HTML'): pagers = ('w3m', 'links', 'lynx', 'less', 'more') else: pagers = ('less', 'more') file_showed = False for pager in pagers: cmd = "%s %s" % (pager, savepath) try: os.system(cmd) except OSError: continue else: file_showed = True break if not file_showed: f = open(savepath) msger.raw(f.read()) f.close() msger.pause() # when proxy needed, make urllib2 follow it proxy = repo.proxy proxy_username = repo.proxy_username proxy_password = repo.proxy_password if not proxy: proxy = get_proxy_for(repo.baseurl[0]) handlers = [] auth_handler = u2.HTTPBasicAuthHandler(u2.HTTPPasswordMgrWithDefaultRealm()) u2opener = None if proxy: if proxy_username: proxy_netloc = urlparse.urlsplit(proxy).netloc if proxy_password: proxy_url = 'http://%s:%s@%s' % (proxy_username, proxy_password, proxy_netloc) else: proxy_url = 'http://%s@%s' % (proxy_username, proxy_netloc) else: proxy_url = proxy proxy_support = u2.ProxyHandler({'http': proxy_url, 'https': proxy_url, 'ftp': proxy_url}) handlers.append(proxy_support) # download all remote files to one temp dir baseurl = None repo_lic_dir = tempfile.mkdtemp(prefix = 'repolic') for url in repo.baseurl: tmphandlers = handlers[:] (scheme, host, path, parm, query, frag) = urlparse.urlparse(url.rstrip('/') + '/') if scheme not in ("http", "https", "ftp", "ftps", "file"): raise CreatorError("Error: invalid url %s" % url) if '@' in host: try: user_pass, host = host.split('@', 1) if ':' in user_pass: user, password = user_pass.split(':', 1) except ValueError, e: raise CreatorError('Bad URL: %s' % url) msger.verbose("adding HTTP auth: %s, XXXXXXXX" %(user)) auth_handler.add_password(None, host, user, password) tmphandlers.append(auth_handler) url = scheme + "://" + host + path + parm + query + frag if tmphandlers: u2opener = u2.build_opener(*tmphandlers) # try to download repo_eula_url = urlparse.urljoin(url, "LICENSE.txt") repo_eula_path = _check_and_download_url( u2opener, repo_eula_url, os.path.join(repo_lic_dir, repo.id + '_LICENSE.txt')) if repo_eula_path: # found baseurl = url break if not baseurl: shutil.rmtree(repo_lic_dir) #cleanup return True # show the license file msger.info('For the software packages in this yum repo:') msger.info(' %s: %s' % (name, baseurl)) msger.info('There is an "End User License Agreement" file that need to be checked.') msger.info('Please read the terms and conditions outlined in it and answer the followed qustions.') msger.pause() _pager_file(repo_eula_path) # Asking for the "Accept/Decline" if not msger.ask('Would you agree to the terms and conditions outlined in the above End User License Agreement?'): msger.warning('Will not install pkgs from this repo.') shutil.rmtree(repo_lic_dir) #cleanup return False # try to find support_info.html for extra infomation repo_info_url = urlparse.urljoin(baseurl, "support_info.html") repo_info_path = _check_and_download_url( u2opener, repo_info_url, os.path.join(repo_lic_dir, repo.id + '_support_info.html')) if repo_info_path: msger.info('There is one more file in the repo for additional support information, please read it') msger.pause() _pager_file(repo_info_path) #cleanup shutil.rmtree(repo_lic_dir) return True ########NEW FILE######## __FILENAME__ = runner #!/usr/bin/python -tt # # Copyright (c) 2011 Intel, Inc. # # This program is free software; you can redistribute it and/or modify it # under the terms of the GNU General Public License as published by the Free # Software Foundation; version 2 of the License # # This program is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY # or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License # for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., 59 # Temple Place - Suite 330, Boston, MA 02111-1307, USA. import os import subprocess from mic import msger def runtool(cmdln_or_args, catch=1): """ wrapper for most of the subprocess calls input: cmdln_or_args: can be both args and cmdln str (shell=True) catch: 0, quitely run 1, only STDOUT 2, only STDERR 3, both STDOUT and STDERR return: (rc, output) if catch==0: the output will always None """ if catch not in (0, 1, 2, 3): # invalid catch selection, will cause exception, that's good return None if isinstance(cmdln_or_args, list): cmd = cmdln_or_args[0] shell = False else: import shlex cmd = shlex.split(cmdln_or_args)[0] shell = True if catch != 3: dev_null = os.open("/dev/null", os.O_WRONLY) if catch == 0: sout = dev_null serr = dev_null elif catch == 1: sout = subprocess.PIPE serr = dev_null elif catch == 2: sout = dev_null serr = subprocess.PIPE elif catch == 3: sout = subprocess.PIPE serr = subprocess.STDOUT try: p = subprocess.Popen(cmdln_or_args, stdout=sout, stderr=serr, shell=shell) (sout, serr) = p.communicate() # combine stdout and stderr, filter None out out = ''.join(filter(None, [sout, serr])) except OSError, e: if e.errno == 2: # [Errno 2] No such file or directory msger.error('Cannot run command: %s, lost dependency?' % cmd) else: raise # relay finally: if catch != 3: os.close(dev_null) return (p.returncode, out) def show(cmdln_or_args): # show all the message using msger.verbose rc, out = runtool(cmdln_or_args, catch=3) if isinstance(cmdln_or_args, list): cmd = ' '.join(cmdln_or_args) else: cmd = cmdln_or_args msg = 'running command: "%s"' % cmd if out: out = out.strip() if out: msg += ', with output::' msg += '\n +----------------' for line in out.splitlines(): msg += '\n | %s' % line msg += '\n +----------------' msger.verbose(msg) return rc def outs(cmdln_or_args, catch=1): # get the outputs of tools return runtool(cmdln_or_args, catch)[1].strip() def quiet(cmdln_or_args): return runtool(cmdln_or_args, catch=0)[0] ########NEW FILE######## __FILENAME__ = __version__ VERSION = "0.14" ########NEW FILE######## __FILENAME__ = oepydevshell-internal #!/usr/bin/env python import os import sys import time import select import fcntl import termios import readline import signal def nonblockingfd(fd): fcntl.fcntl(fd, fcntl.F_SETFL, fcntl.fcntl(fd, fcntl.F_GETFL) | os.O_NONBLOCK) def echonocbreak(fd): old = termios.tcgetattr(fd) old[3] = old[3] | termios.ECHO | termios.ICANON termios.tcsetattr(fd, termios.TCSADRAIN, old) def cbreaknoecho(fd): old = termios.tcgetattr(fd) old[3] = old[3] &~ termios.ECHO &~ termios.ICANON termios.tcsetattr(fd, termios.TCSADRAIN, old) if len(sys.argv) != 3: print("Incorrect parameters") sys.exit(1) pty = open(sys.argv[1], "w+b", 0) parent = int(sys.argv[2]) # Don't buffer output by line endings sys.stdout = os.fdopen(sys.stdout.fileno(), 'w', 0) sys.stdin = os.fdopen(sys.stdin.fileno(), 'r', 0) nonblockingfd(pty) nonblockingfd(sys.stdin) histfile = os.path.expanduser("~/.oedevpyshell-history") readline.parse_and_bind("tab: complete") try: readline.read_history_file(histfile) except IOError: pass try: i = "" o = "" # Need cbreak/noecho whilst in select so we trigger on any keypress cbreaknoecho(sys.stdin.fileno()) # Send our PID to the other end so they can kill us. pty.write(str(os.getpid()) + "\n") while True: try: writers = [] if i: writers.append(sys.stdout) (ready, _, _) = select.select([pty, sys.stdin], writers , [], 0) try: if pty in ready: i = i + pty.read() if i: # Write a page at a time to avoid overflowing output # d.keys() is a good way to do that sys.stdout.write(i[:4096]) i = i[4096:] if sys.stdin in ready: echonocbreak(sys.stdin.fileno()) o = raw_input() cbreaknoecho(sys.stdin.fileno()) pty.write(o + "\n") except (IOError, OSError) as e: if e.errno == 11: continue if e.errno == 5: sys.exit(0) raise except EOFError: sys.exit(0) except KeyboardInterrupt: os.kill(parent, signal.SIGINT) except SystemExit: pass except Exception as e: import traceback print("Exception in oepydehshell-internal: " + str(e)) traceback.print_exc() time.sleep(5) finally: readline.write_history_file(histfile) ########NEW FILE######## __FILENAME__ = opkg-query-helper #!/usr/bin/env python # OpenEmbedded opkg query helper utility # # Written by: Paul Eggleton <paul.eggleton@linux.intel.com> # # Copyright 2012 Intel Corporation # # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License version 2 as # published by the Free Software Foundation. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., # 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA. # # import sys import fileinput import re archmode = False filemode = False vermode = False args = [] for arg in sys.argv[1:]: if arg == '-a': archmode = True elif arg == '-f': filemode = True elif arg == '-v': vermode = True else: args.append(arg) # Regex for removing version specs after dependency items verregex = re.compile(' \([=<>]* [^ )]*\)') pkg = "" ver = "" for line in fileinput.input(args): line = line.rstrip() if ': ' in line: if line.startswith("Package:"): pkg = line.split(": ")[1] ver = "" else: if archmode: if line.startswith("Architecture:"): arch = line.split(": ")[1] print("%s %s" % (pkg,arch)) elif filemode: if line.startswith("Version:"): ver = line.split(": ")[1] elif line.startswith("Architecture:"): arch = line.split(": ")[1] print("%s %s_%s_%s.ipk %s" % (pkg,pkg,ver,arch,arch)) elif vermode: if line.startswith("Version:"): ver = line.split(": ")[1] elif line.startswith("Architecture:"): arch = line.split(": ")[1] print("%s %s %s" % (pkg,arch,ver)) else: if line.startswith("Depends:"): depval = line.split(": ")[1] deps = depval.split(", ") for dep in deps: dep = verregex.sub('', dep) print("%s|%s" % (pkg,dep)) elif line.startswith("Recommends:"): recval = line.split(": ")[1] recs = recval.split(", ") for rec in recs: rec = verregex.sub('', rec) print("%s|%s [REC]" % (pkg, rec)) ########NEW FILE######## __FILENAME__ = batch # This file is part of pybootchartgui. # pybootchartgui is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # pybootchartgui is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # You should have received a copy of the GNU General Public License # along with pybootchartgui. If not, see <http://www.gnu.org/licenses/>. import cairo from . import draw from .draw import RenderOptions def render(writer, trace, app_options, filename): handlers = { "png": (lambda w, h: cairo.ImageSurface(cairo.FORMAT_ARGB32, w, h), \ lambda sfc: sfc.write_to_png(filename)), "pdf": (lambda w, h: cairo.PDFSurface(filename, w, h), lambda sfc: 0), "svg": (lambda w, h: cairo.SVGSurface(filename, w, h), lambda sfc: 0) } if app_options.format is None: fmt = filename.rsplit('.', 1)[1] else: fmt = app_options.format if not (fmt in handlers): writer.error ("Unknown format '%s'." % fmt) return 10 make_surface, write_surface = handlers[fmt] options = RenderOptions (app_options) (w, h) = draw.extents (options, 1.0, trace) w = max (w, draw.MIN_IMG_W) surface = make_surface (w, h) ctx = cairo.Context (surface) draw.render (ctx, options, 1.0, trace) write_surface (surface) writer.status ("bootchart written to '%s'" % filename) ########NEW FILE######## __FILENAME__ = draw # This file is part of pybootchartgui. # pybootchartgui is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # pybootchartgui is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # You should have received a copy of the GNU General Public License # along with pybootchartgui. If not, see <http://www.gnu.org/licenses/>. import cairo import math import re import random import colorsys from operator import itemgetter class RenderOptions: def __init__(self, app_options): # should we render a cumulative CPU time chart self.cumulative = True self.charts = True self.kernel_only = False self.app_options = app_options def proc_tree (self, trace): if self.kernel_only: return trace.kernel_tree else: return trace.proc_tree # Process tree background color. BACK_COLOR = (1.0, 1.0, 1.0, 1.0) WHITE = (1.0, 1.0, 1.0, 1.0) # Process tree border color. BORDER_COLOR = (0.63, 0.63, 0.63, 1.0) # Second tick line color. TICK_COLOR = (0.92, 0.92, 0.92, 1.0) # 5-second tick line color. TICK_COLOR_BOLD = (0.86, 0.86, 0.86, 1.0) # Annotation colour ANNOTATION_COLOR = (0.63, 0.0, 0.0, 0.5) # Text color. TEXT_COLOR = (0.0, 0.0, 0.0, 1.0) # Font family FONT_NAME = "Bitstream Vera Sans" # Title text font. TITLE_FONT_SIZE = 18 # Default text font. TEXT_FONT_SIZE = 12 # Axis label font. AXIS_FONT_SIZE = 11 # Legend font. LEGEND_FONT_SIZE = 12 # CPU load chart color. CPU_COLOR = (0.40, 0.55, 0.70, 1.0) # IO wait chart color. IO_COLOR = (0.76, 0.48, 0.48, 0.5) # Disk throughput color. DISK_TPUT_COLOR = (0.20, 0.71, 0.20, 1.0) # CPU load chart color. FILE_OPEN_COLOR = (0.20, 0.71, 0.71, 1.0) # Mem cached color MEM_CACHED_COLOR = CPU_COLOR # Mem used color MEM_USED_COLOR = IO_COLOR # Buffers color MEM_BUFFERS_COLOR = (0.4, 0.4, 0.4, 0.3) # Swap color MEM_SWAP_COLOR = DISK_TPUT_COLOR # Process border color. PROC_BORDER_COLOR = (0.71, 0.71, 0.71, 1.0) # Waiting process color. PROC_COLOR_D = (0.76, 0.48, 0.48, 0.5) # Running process color. PROC_COLOR_R = CPU_COLOR # Sleeping process color. PROC_COLOR_S = (0.94, 0.94, 0.94, 1.0) # Stopped process color. PROC_COLOR_T = (0.94, 0.50, 0.50, 1.0) # Zombie process color. PROC_COLOR_Z = (0.71, 0.71, 0.71, 1.0) # Dead process color. PROC_COLOR_X = (0.71, 0.71, 0.71, 0.125) # Paging process color. PROC_COLOR_W = (0.71, 0.71, 0.71, 0.125) # Process label color. PROC_TEXT_COLOR = (0.19, 0.19, 0.19, 1.0) # Process label font. PROC_TEXT_FONT_SIZE = 12 # Signature color. SIG_COLOR = (0.0, 0.0, 0.0, 0.3125) # Signature font. SIG_FONT_SIZE = 14 # Signature text. SIGNATURE = "http://github.com/mmeeks/bootchart" # Process dependency line color. DEP_COLOR = (0.75, 0.75, 0.75, 1.0) # Process dependency line stroke. DEP_STROKE = 1.0 # Process description date format. DESC_TIME_FORMAT = "mm:ss.SSS" # Cumulative coloring bits HSV_MAX_MOD = 31 HSV_STEP = 7 # Configure task color TASK_COLOR_CONFIGURE = (1.0, 1.0, 0.00, 1.0) # Compile task color. TASK_COLOR_COMPILE = (0.0, 1.00, 0.00, 1.0) # Install task color TASK_COLOR_INSTALL = (1.0, 0.00, 1.00, 1.0) # Sysroot task color TASK_COLOR_SYSROOT = (0.0, 0.00, 1.00, 1.0) # Package task color TASK_COLOR_PACKAGE = (0.0, 1.00, 1.00, 1.0) # Package Write RPM/DEB/IPK task color TASK_COLOR_PACKAGE_WRITE = (0.0, 0.50, 0.50, 1.0) # Process states STATE_UNDEFINED = 0 STATE_RUNNING = 1 STATE_SLEEPING = 2 STATE_WAITING = 3 STATE_STOPPED = 4 STATE_ZOMBIE = 5 STATE_COLORS = [(0, 0, 0, 0), PROC_COLOR_R, PROC_COLOR_S, PROC_COLOR_D, \ PROC_COLOR_T, PROC_COLOR_Z, PROC_COLOR_X, PROC_COLOR_W] # CumulativeStats Types STAT_TYPE_CPU = 0 STAT_TYPE_IO = 1 # Convert ps process state to an int def get_proc_state(flag): return "RSDTZXW".find(flag) + 1 def draw_text(ctx, text, color, x, y): ctx.set_source_rgba(*color) ctx.move_to(x, y) ctx.show_text(text) def draw_fill_rect(ctx, color, rect): ctx.set_source_rgba(*color) ctx.rectangle(*rect) ctx.fill() def draw_rect(ctx, color, rect): ctx.set_source_rgba(*color) ctx.rectangle(*rect) ctx.stroke() def draw_legend_box(ctx, label, fill_color, x, y, s): draw_fill_rect(ctx, fill_color, (x, y - s, s, s)) draw_rect(ctx, PROC_BORDER_COLOR, (x, y - s, s, s)) draw_text(ctx, label, TEXT_COLOR, x + s + 5, y) def draw_legend_line(ctx, label, fill_color, x, y, s): draw_fill_rect(ctx, fill_color, (x, y - s/2, s + 1, 3)) ctx.arc(x + (s + 1)/2.0, y - (s - 3)/2.0, 2.5, 0, 2.0 * math.pi) ctx.fill() draw_text(ctx, label, TEXT_COLOR, x + s + 5, y) def draw_label_in_box(ctx, color, label, x, y, w, maxx): label_w = ctx.text_extents(label)[2] label_x = x + w / 2 - label_w / 2 if label_w + 10 > w: label_x = x + w + 5 if label_x + label_w > maxx: label_x = x - label_w - 5 draw_text(ctx, label, color, label_x, y) def draw_sec_labels(ctx, options, rect, sec_w, nsecs): ctx.set_font_size(AXIS_FONT_SIZE) prev_x = 0 for i in range(0, rect[2] + 1, sec_w): if ((i / sec_w) % nsecs == 0) : if options.app_options.as_minutes : label = "%.1f" % (i / sec_w / 60.0) else : label = "%d" % (i / sec_w) label_w = ctx.text_extents(label)[2] x = rect[0] + i - label_w/2 if x >= prev_x: draw_text(ctx, label, TEXT_COLOR, x, rect[1] - 2) prev_x = x + label_w def draw_box_ticks(ctx, rect, sec_w): draw_rect(ctx, BORDER_COLOR, tuple(rect)) ctx.set_line_cap(cairo.LINE_CAP_SQUARE) for i in range(sec_w, rect[2] + 1, sec_w): if ((i / sec_w) % 10 == 0) : ctx.set_line_width(1.5) elif sec_w < 5 : continue else : ctx.set_line_width(1.0) if ((i / sec_w) % 30 == 0) : ctx.set_source_rgba(*TICK_COLOR_BOLD) else : ctx.set_source_rgba(*TICK_COLOR) ctx.move_to(rect[0] + i, rect[1] + 1) ctx.line_to(rect[0] + i, rect[1] + rect[3] - 1) ctx.stroke() ctx.set_line_width(1.0) ctx.set_line_cap(cairo.LINE_CAP_BUTT) def draw_annotations(ctx, proc_tree, times, rect): ctx.set_line_cap(cairo.LINE_CAP_SQUARE) ctx.set_source_rgba(*ANNOTATION_COLOR) ctx.set_dash([4, 4]) for time in times: if time is not None: x = ((time - proc_tree.start_time) * rect[2] / proc_tree.duration) ctx.move_to(rect[0] + x, rect[1] + 1) ctx.line_to(rect[0] + x, rect[1] + rect[3] - 1) ctx.stroke() ctx.set_line_cap(cairo.LINE_CAP_BUTT) ctx.set_dash([]) def draw_chart(ctx, color, fill, chart_bounds, data, proc_tree, data_range): ctx.set_line_width(0.5) x_shift = proc_tree.start_time def transform_point_coords(point, x_base, y_base, \ xscale, yscale, x_trans, y_trans): x = (point[0] - x_base) * xscale + x_trans y = (point[1] - y_base) * -yscale + y_trans + chart_bounds[3] return x, y max_x = max (x for (x, y) in data) max_y = max (y for (x, y) in data) # avoid divide by zero if max_y == 0: max_y = 1.0 xscale = float (chart_bounds[2]) / max_x # If data_range is given, scale the chart so that the value range in # data_range matches the chart bounds exactly. # Otherwise, scale so that the actual data matches the chart bounds. if data_range: yscale = float(chart_bounds[3]) / (data_range[1] - data_range[0]) ybase = data_range[0] else: yscale = float(chart_bounds[3]) / max_y ybase = 0 first = transform_point_coords (data[0], x_shift, ybase, xscale, yscale, \ chart_bounds[0], chart_bounds[1]) last = transform_point_coords (data[-1], x_shift, ybase, xscale, yscale, \ chart_bounds[0], chart_bounds[1]) ctx.set_source_rgba(*color) ctx.move_to(*first) for point in data: x, y = transform_point_coords (point, x_shift, ybase, xscale, yscale, \ chart_bounds[0], chart_bounds[1]) ctx.line_to(x, y) if fill: ctx.stroke_preserve() ctx.line_to(last[0], chart_bounds[1]+chart_bounds[3]) ctx.line_to(first[0], chart_bounds[1]+chart_bounds[3]) ctx.line_to(first[0], first[1]) ctx.fill() else: ctx.stroke() ctx.set_line_width(1.0) bar_h = 55 meminfo_bar_h = 2 * bar_h header_h = 60 # offsets off_x, off_y = 220, 10 sec_w_base = 1 # the width of a second proc_h = 16 # the height of a process leg_s = 10 MIN_IMG_W = 800 CUML_HEIGHT = 2000 # Increased value to accomodate CPU and I/O Graphs OPTIONS = None def extents(options, xscale, trace): start = min(trace.start.keys()) end = start processes = 0 for proc in trace.processes: if not options.app_options.show_all and \ trace.processes[proc][1] - trace.processes[proc][0] < options.app_options.mintime: continue if trace.processes[proc][1] > end: end = trace.processes[proc][1] processes += 1 if trace.min is not None and trace.max is not None: start = trace.min end = trace.max w = int ((end - start) * sec_w_base * xscale) + 2 * off_x h = proc_h * processes + header_h + 2 * off_y return (w, h) def clip_visible(clip, rect): xmax = max (clip[0], rect[0]) ymax = max (clip[1], rect[1]) xmin = min (clip[0] + clip[2], rect[0] + rect[2]) ymin = min (clip[1] + clip[3], rect[1] + rect[3]) return (xmin > xmax and ymin > ymax) def render_charts(ctx, options, clip, trace, curr_y, w, h, sec_w): proc_tree = options.proc_tree(trace) # render bar legend ctx.set_font_size(LEGEND_FONT_SIZE) draw_legend_box(ctx, "CPU (user+sys)", CPU_COLOR, off_x, curr_y+20, leg_s) draw_legend_box(ctx, "I/O (wait)", IO_COLOR, off_x + 120, curr_y+20, leg_s) # render I/O wait chart_rect = (off_x, curr_y+30, w, bar_h) if clip_visible (clip, chart_rect): draw_box_ticks (ctx, chart_rect, sec_w) draw_annotations (ctx, proc_tree, trace.times, chart_rect) draw_chart (ctx, IO_COLOR, True, chart_rect, \ [(sample.time, sample.user + sample.sys + sample.io) for sample in trace.cpu_stats], \ proc_tree, None) # render CPU load draw_chart (ctx, CPU_COLOR, True, chart_rect, \ [(sample.time, sample.user + sample.sys) for sample in trace.cpu_stats], \ proc_tree, None) curr_y = curr_y + 30 + bar_h # render second chart draw_legend_line(ctx, "Disk throughput", DISK_TPUT_COLOR, off_x, curr_y+20, leg_s) draw_legend_box(ctx, "Disk utilization", IO_COLOR, off_x + 120, curr_y+20, leg_s) # render I/O utilization chart_rect = (off_x, curr_y+30, w, bar_h) if clip_visible (clip, chart_rect): draw_box_ticks (ctx, chart_rect, sec_w) draw_annotations (ctx, proc_tree, trace.times, chart_rect) draw_chart (ctx, IO_COLOR, True, chart_rect, \ [(sample.time, sample.util) for sample in trace.disk_stats], \ proc_tree, None) # render disk throughput max_sample = max (trace.disk_stats, key = lambda s: s.tput) if clip_visible (clip, chart_rect): draw_chart (ctx, DISK_TPUT_COLOR, False, chart_rect, \ [(sample.time, sample.tput) for sample in trace.disk_stats], \ proc_tree, None) pos_x = off_x + ((max_sample.time - proc_tree.start_time) * w / proc_tree.duration) shift_x, shift_y = -20, 20 if (pos_x < off_x + 245): shift_x, shift_y = 5, 40 label = "%dMB/s" % round ((max_sample.tput) / 1024.0) draw_text (ctx, label, DISK_TPUT_COLOR, pos_x + shift_x, curr_y + shift_y) curr_y = curr_y + 30 + bar_h # render mem usage chart_rect = (off_x, curr_y+30, w, meminfo_bar_h) mem_stats = trace.mem_stats if mem_stats and clip_visible (clip, chart_rect): mem_scale = max(sample.records['MemTotal'] - sample.records['MemFree'] for sample in mem_stats) draw_legend_box(ctx, "Mem cached (scale: %u MiB)" % (float(mem_scale) / 1024), MEM_CACHED_COLOR, off_x, curr_y+20, leg_s) draw_legend_box(ctx, "Used", MEM_USED_COLOR, off_x + 240, curr_y+20, leg_s) draw_legend_box(ctx, "Buffers", MEM_BUFFERS_COLOR, off_x + 360, curr_y+20, leg_s) draw_legend_line(ctx, "Swap (scale: %u MiB)" % max([(sample.records['SwapTotal'] - sample.records['SwapFree'])/1024 for sample in mem_stats]), \ MEM_SWAP_COLOR, off_x + 480, curr_y+20, leg_s) draw_box_ticks(ctx, chart_rect, sec_w) draw_annotations(ctx, proc_tree, trace.times, chart_rect) draw_chart(ctx, MEM_BUFFERS_COLOR, True, chart_rect, \ [(sample.time, sample.records['MemTotal'] - sample.records['MemFree']) for sample in trace.mem_stats], \ proc_tree, [0, mem_scale]) draw_chart(ctx, MEM_USED_COLOR, True, chart_rect, \ [(sample.time, sample.records['MemTotal'] - sample.records['MemFree'] - sample.records['Buffers']) for sample in mem_stats], \ proc_tree, [0, mem_scale]) draw_chart(ctx, MEM_CACHED_COLOR, True, chart_rect, \ [(sample.time, sample.records['Cached']) for sample in mem_stats], \ proc_tree, [0, mem_scale]) draw_chart(ctx, MEM_SWAP_COLOR, False, chart_rect, \ [(sample.time, float(sample.records['SwapTotal'] - sample.records['SwapFree'])) for sample in mem_stats], \ proc_tree, None) curr_y = curr_y + meminfo_bar_h return curr_y def render_processes_chart(ctx, options, trace, curr_y, w, h, sec_w): chart_rect = [off_x, curr_y+header_h, w, h - 2 * off_y - (curr_y+header_h) + proc_h] draw_legend_box (ctx, "Configure", \ TASK_COLOR_CONFIGURE, off_x , curr_y + 45, leg_s) draw_legend_box (ctx, "Compile", \ TASK_COLOR_COMPILE, off_x+120, curr_y + 45, leg_s) draw_legend_box (ctx, "Install", \ TASK_COLOR_INSTALL, off_x+240, curr_y + 45, leg_s) draw_legend_box (ctx, "Populate Sysroot", \ TASK_COLOR_SYSROOT, off_x+360, curr_y + 45, leg_s) draw_legend_box (ctx, "Package", \ TASK_COLOR_PACKAGE, off_x+480, curr_y + 45, leg_s) draw_legend_box (ctx, "Package Write", TASK_COLOR_PACKAGE_WRITE, off_x+600, curr_y + 45, leg_s) ctx.set_font_size(PROC_TEXT_FONT_SIZE) draw_box_ticks(ctx, chart_rect, sec_w) draw_sec_labels(ctx, options, chart_rect, sec_w, 30) y = curr_y+header_h offset = trace.min or min(trace.start.keys()) for s in sorted(trace.start.keys()): for val in sorted(trace.start[s]): if not options.app_options.show_all and \ trace.processes[val][1] - s < options.app_options.mintime: continue task = val.split(":")[1] #print val #print trace.processes[val][1] #print s x = chart_rect[0] + (s - offset) * sec_w w = ((trace.processes[val][1] - s) * sec_w) #print "proc at %s %s %s %s" % (x, y, w, proc_h) col = None if task == "do_compile": col = TASK_COLOR_COMPILE elif task == "do_configure": col = TASK_COLOR_CONFIGURE elif task == "do_install": col = TASK_COLOR_INSTALL elif task == "do_populate_sysroot": col = TASK_COLOR_SYSROOT elif task == "do_package": col = TASK_COLOR_PACKAGE elif task == "do_package_write_rpm" or \ task == "do_package_write_deb" or \ task == "do_package_write_ipk": col = TASK_COLOR_PACKAGE_WRITE else: col = WHITE if col: draw_fill_rect(ctx, col, (x, y, w, proc_h)) draw_rect(ctx, PROC_BORDER_COLOR, (x, y, w, proc_h)) draw_label_in_box(ctx, PROC_TEXT_COLOR, val, x, y + proc_h - 4, w, proc_h) y = y + proc_h return curr_y # # Render the chart. # def render(ctx, options, xscale, trace): (w, h) = extents (options, xscale, trace) global OPTIONS OPTIONS = options.app_options # x, y, w, h clip = ctx.clip_extents() sec_w = int (xscale * sec_w_base) ctx.set_line_width(1.0) ctx.select_font_face(FONT_NAME) draw_fill_rect(ctx, WHITE, (0, 0, max(w, MIN_IMG_W), h)) w -= 2*off_x curr_y = off_y; curr_y = render_processes_chart (ctx, options, trace, curr_y, w, h, sec_w) return proc_tree = options.proc_tree (trace) # draw the title and headers if proc_tree.idle: duration = proc_tree.idle else: duration = proc_tree.duration if not options.kernel_only: curr_y = draw_header (ctx, trace.headers, duration) else: curr_y = off_y; if options.charts: curr_y = render_charts (ctx, options, clip, trace, curr_y, w, h, sec_w) # draw process boxes proc_height = h if proc_tree.taskstats and options.cumulative: proc_height -= CUML_HEIGHT draw_process_bar_chart(ctx, clip, options, proc_tree, trace.times, curr_y, w, proc_height, sec_w) curr_y = proc_height ctx.set_font_size(SIG_FONT_SIZE) draw_text(ctx, SIGNATURE, SIG_COLOR, off_x + 5, proc_height - 8) # draw a cumulative CPU-time-per-process graph if proc_tree.taskstats and options.cumulative: cuml_rect = (off_x, curr_y + off_y, w, CUML_HEIGHT/2 - off_y * 2) if clip_visible (clip, cuml_rect): draw_cuml_graph(ctx, proc_tree, cuml_rect, duration, sec_w, STAT_TYPE_CPU) # draw a cumulative I/O-time-per-process graph if proc_tree.taskstats and options.cumulative: cuml_rect = (off_x, curr_y + off_y * 100, w, CUML_HEIGHT/2 - off_y * 2) if clip_visible (clip, cuml_rect): draw_cuml_graph(ctx, proc_tree, cuml_rect, duration, sec_w, STAT_TYPE_IO) def draw_process_bar_chart(ctx, clip, options, proc_tree, times, curr_y, w, h, sec_w): header_size = 0 if not options.kernel_only: draw_legend_box (ctx, "Running (%cpu)", PROC_COLOR_R, off_x , curr_y + 45, leg_s) draw_legend_box (ctx, "Unint.sleep (I/O)", PROC_COLOR_D, off_x+120, curr_y + 45, leg_s) draw_legend_box (ctx, "Sleeping", PROC_COLOR_S, off_x+240, curr_y + 45, leg_s) draw_legend_box (ctx, "Zombie", PROC_COLOR_Z, off_x+360, curr_y + 45, leg_s) header_size = 45 chart_rect = [off_x, curr_y + header_size + 15, w, h - 2 * off_y - (curr_y + header_size + 15) + proc_h] ctx.set_font_size (PROC_TEXT_FONT_SIZE) draw_box_ticks (ctx, chart_rect, sec_w) if sec_w > 100: nsec = 1 else: nsec = 5 draw_sec_labels (ctx, options, chart_rect, sec_w, nsec) draw_annotations (ctx, proc_tree, times, chart_rect) y = curr_y + 60 for root in proc_tree.process_tree: draw_processes_recursively(ctx, root, proc_tree, y, proc_h, chart_rect, clip) y = y + proc_h * proc_tree.num_nodes([root]) def draw_header (ctx, headers, duration): toshow = [ ('system.uname', 'uname', lambda s: s), ('system.release', 'release', lambda s: s), ('system.cpu', 'CPU', lambda s: re.sub('model name\s*:\s*', '', s, 1)), ('system.kernel.options', 'kernel options', lambda s: s), ] header_y = ctx.font_extents()[2] + 10 ctx.set_font_size(TITLE_FONT_SIZE) draw_text(ctx, headers['title'], TEXT_COLOR, off_x, header_y) ctx.set_font_size(TEXT_FONT_SIZE) for (headerkey, headertitle, mangle) in toshow: header_y += ctx.font_extents()[2] if headerkey in headers: value = headers.get(headerkey) else: value = "" txt = headertitle + ': ' + mangle(value) draw_text(ctx, txt, TEXT_COLOR, off_x, header_y) dur = duration / 100.0 txt = 'time : %02d:%05.2f' % (math.floor(dur/60), dur - 60 * math.floor(dur/60)) if headers.get('system.maxpid') is not None: txt = txt + ' max pid: %s' % (headers.get('system.maxpid')) header_y += ctx.font_extents()[2] draw_text (ctx, txt, TEXT_COLOR, off_x, header_y) return header_y def draw_processes_recursively(ctx, proc, proc_tree, y, proc_h, rect, clip) : x = rect[0] + ((proc.start_time - proc_tree.start_time) * rect[2] / proc_tree.duration) w = ((proc.duration) * rect[2] / proc_tree.duration) draw_process_activity_colors(ctx, proc, proc_tree, x, y, w, proc_h, rect, clip) draw_rect(ctx, PROC_BORDER_COLOR, (x, y, w, proc_h)) ipid = int(proc.pid) if not OPTIONS.show_all: cmdString = proc.cmd else: cmdString = '' if (OPTIONS.show_pid or OPTIONS.show_all) and ipid is not 0: cmdString = cmdString + " [" + str(ipid // 1000) + "]" if OPTIONS.show_all: if proc.args: cmdString = cmdString + " '" + "' '".join(proc.args) + "'" else: cmdString = cmdString + " " + proc.exe draw_label_in_box(ctx, PROC_TEXT_COLOR, cmdString, x, y + proc_h - 4, w, rect[0] + rect[2]) next_y = y + proc_h for child in proc.child_list: if next_y > clip[1] + clip[3]: break child_x, child_y = draw_processes_recursively(ctx, child, proc_tree, next_y, proc_h, rect, clip) draw_process_connecting_lines(ctx, x, y, child_x, child_y, proc_h) next_y = next_y + proc_h * proc_tree.num_nodes([child]) return x, y def draw_process_activity_colors(ctx, proc, proc_tree, x, y, w, proc_h, rect, clip): if y > clip[1] + clip[3] or y + proc_h + 2 < clip[1]: return draw_fill_rect(ctx, PROC_COLOR_S, (x, y, w, proc_h)) last_tx = -1 for sample in proc.samples : tx = rect[0] + round(((sample.time - proc_tree.start_time) * rect[2] / proc_tree.duration)) # samples are sorted chronologically if tx < clip[0]: continue if tx > clip[0] + clip[2]: break tw = round(proc_tree.sample_period * rect[2] / float(proc_tree.duration)) if last_tx != -1 and abs(last_tx - tx) <= tw: tw -= last_tx - tx tx = last_tx tw = max (tw, 1) # nice to see at least something last_tx = tx + tw state = get_proc_state( sample.state ) color = STATE_COLORS[state] if state == STATE_RUNNING: alpha = min (sample.cpu_sample.user + sample.cpu_sample.sys, 1.0) color = tuple(list(PROC_COLOR_R[0:3]) + [alpha]) # print "render time %d [ tx %d tw %d ], sample state %s color %s alpha %g" % (sample.time, tx, tw, state, color, alpha) elif state == STATE_SLEEPING: continue draw_fill_rect(ctx, color, (tx, y, tw, proc_h)) def draw_process_connecting_lines(ctx, px, py, x, y, proc_h): ctx.set_source_rgba(*DEP_COLOR) ctx.set_dash([2, 2]) if abs(px - x) < 3: dep_off_x = 3 dep_off_y = proc_h / 4 ctx.move_to(x, y + proc_h / 2) ctx.line_to(px - dep_off_x, y + proc_h / 2) ctx.line_to(px - dep_off_x, py - dep_off_y) ctx.line_to(px, py - dep_off_y) else: ctx.move_to(x, y + proc_h / 2) ctx.line_to(px, y + proc_h / 2) ctx.line_to(px, py) ctx.stroke() ctx.set_dash([]) # elide the bootchart collector - it is quite distorting def elide_bootchart(proc): return proc.cmd == 'bootchartd' or proc.cmd == 'bootchart-colle' class CumlSample: def __init__(self, proc): self.cmd = proc.cmd self.samples = [] self.merge_samples (proc) self.color = None def merge_samples(self, proc): self.samples.extend (proc.samples) self.samples.sort (key = lambda p: p.time) def next(self): global palette_idx palette_idx += HSV_STEP return palette_idx def get_color(self): if self.color is None: i = self.next() % HSV_MAX_MOD h = 0.0 if i is not 0: h = (1.0 * i) / HSV_MAX_MOD s = 0.5 v = 1.0 c = colorsys.hsv_to_rgb (h, s, v) self.color = (c[0], c[1], c[2], 1.0) return self.color def draw_cuml_graph(ctx, proc_tree, chart_bounds, duration, sec_w, stat_type): global palette_idx palette_idx = 0 time_hash = {} total_time = 0.0 m_proc_list = {} if stat_type is STAT_TYPE_CPU: sample_value = 'cpu' else: sample_value = 'io' for proc in proc_tree.process_list: if elide_bootchart(proc): continue for sample in proc.samples: total_time += getattr(sample.cpu_sample, sample_value) if not sample.time in time_hash: time_hash[sample.time] = 1 # merge pids with the same cmd if not proc.cmd in m_proc_list: m_proc_list[proc.cmd] = CumlSample (proc) continue s = m_proc_list[proc.cmd] s.merge_samples (proc) # all the sample times times = sorted(time_hash) if len (times) < 2: print("degenerate boot chart") return pix_per_ns = chart_bounds[3] / total_time # print "total time: %g pix-per-ns %g" % (total_time, pix_per_ns) # FIXME: we have duplicates in the process list too [!] - why !? # Render bottom up, left to right below = {} for time in times: below[time] = chart_bounds[1] + chart_bounds[3] # same colors each time we render random.seed (0) ctx.set_line_width(1) legends = [] labels = [] # render each pid in order for cs in m_proc_list.values(): row = {} cuml = 0.0 # print "pid : %s -> %g samples %d" % (proc.cmd, cuml, len (cs.samples)) for sample in cs.samples: cuml += getattr(sample.cpu_sample, sample_value) row[sample.time] = cuml process_total_time = cuml # hide really tiny processes if cuml * pix_per_ns <= 2: continue last_time = times[0] y = last_below = below[last_time] last_cuml = cuml = 0.0 ctx.set_source_rgba(*cs.get_color()) for time in times: render_seg = False # did the underlying trend increase ? if below[time] != last_below: last_below = below[last_time] last_cuml = cuml render_seg = True # did we move up a pixel increase ? if time in row: nc = round (row[time] * pix_per_ns) if nc != cuml: last_cuml = cuml cuml = nc render_seg = True # if last_cuml > cuml: # assert fail ... - un-sorted process samples # draw the trailing rectangle from the last time to # before now, at the height of the last segment. if render_seg: w = math.ceil ((time - last_time) * chart_bounds[2] / proc_tree.duration) + 1 x = chart_bounds[0] + round((last_time - proc_tree.start_time) * chart_bounds[2] / proc_tree.duration) ctx.rectangle (x, below[last_time] - last_cuml, w, last_cuml) ctx.fill() # ctx.stroke() last_time = time y = below [time] - cuml row[time] = y # render the last segment x = chart_bounds[0] + round((last_time - proc_tree.start_time) * chart_bounds[2] / proc_tree.duration) y = below[last_time] - cuml ctx.rectangle (x, y, chart_bounds[2] - x, cuml) ctx.fill() # ctx.stroke() # render legend if it will fit if cuml > 8: label = cs.cmd extnts = ctx.text_extents(label) label_w = extnts[2] label_h = extnts[3] # print "Text extents %g by %g" % (label_w, label_h) labels.append((label, chart_bounds[0] + chart_bounds[2] - label_w - off_x * 2, y + (cuml + label_h) / 2)) if cs in legends: print("ARGH - duplicate process in list !") legends.append ((cs, process_total_time)) below = row # render grid-lines over the top draw_box_ticks(ctx, chart_bounds, sec_w) # render labels for l in labels: draw_text(ctx, l[0], TEXT_COLOR, l[1], l[2]) # Render legends font_height = 20 label_width = 300 LEGENDS_PER_COL = 15 LEGENDS_TOTAL = 45 ctx.set_font_size (TITLE_FONT_SIZE) dur_secs = duration / 100 cpu_secs = total_time / 1000000000 # misleading - with multiple CPUs ... # idle = ((dur_secs - cpu_secs) / dur_secs) * 100.0 if stat_type is STAT_TYPE_CPU: label = "Cumulative CPU usage, by process; total CPU: " \ " %.5g(s) time: %.3g(s)" % (cpu_secs, dur_secs) else: label = "Cumulative I/O usage, by process; total I/O: " \ " %.5g(s) time: %.3g(s)" % (cpu_secs, dur_secs) draw_text(ctx, label, TEXT_COLOR, chart_bounds[0] + off_x, chart_bounds[1] + font_height) i = 0 legends = sorted(legends, key=itemgetter(1), reverse=True) ctx.set_font_size(TEXT_FONT_SIZE) for t in legends: cs = t[0] time = t[1] x = chart_bounds[0] + off_x + int (i/LEGENDS_PER_COL) * label_width y = chart_bounds[1] + font_height * ((i % LEGENDS_PER_COL) + 2) str = "%s - %.0f(ms) (%2.2f%%)" % (cs.cmd, time/1000000, (time/total_time) * 100.0) draw_legend_box(ctx, str, cs.color, x, y, leg_s) i = i + 1 if i >= LEGENDS_TOTAL: break ########NEW FILE######## __FILENAME__ = gui # This file is part of pybootchartgui. # pybootchartgui is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # pybootchartgui is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # You should have received a copy of the GNU General Public License # along with pybootchartgui. If not, see <http://www.gnu.org/licenses/>. import gobject import gtk import gtk.gdk import gtk.keysyms from . import draw from .draw import RenderOptions class PyBootchartWidget(gtk.DrawingArea): __gsignals__ = { 'expose-event': 'override', 'clicked' : (gobject.SIGNAL_RUN_LAST, gobject.TYPE_NONE, (gobject.TYPE_STRING, gtk.gdk.Event)), 'position-changed' : (gobject.SIGNAL_RUN_LAST, gobject.TYPE_NONE, (gobject.TYPE_INT, gobject.TYPE_INT)), 'set-scroll-adjustments' : (gobject.SIGNAL_RUN_LAST, gobject.TYPE_NONE, (gtk.Adjustment, gtk.Adjustment)) } def __init__(self, trace, options, xscale): gtk.DrawingArea.__init__(self) self.trace = trace self.options = options self.set_flags(gtk.CAN_FOCUS) self.add_events(gtk.gdk.BUTTON_PRESS_MASK | gtk.gdk.BUTTON_RELEASE_MASK) self.connect("button-press-event", self.on_area_button_press) self.connect("button-release-event", self.on_area_button_release) self.add_events(gtk.gdk.POINTER_MOTION_MASK | gtk.gdk.POINTER_MOTION_HINT_MASK | gtk.gdk.BUTTON_RELEASE_MASK) self.connect("motion-notify-event", self.on_area_motion_notify) self.connect("scroll-event", self.on_area_scroll_event) self.connect('key-press-event', self.on_key_press_event) self.connect('set-scroll-adjustments', self.on_set_scroll_adjustments) self.connect("size-allocate", self.on_allocation_size_changed) self.connect("position-changed", self.on_position_changed) self.zoom_ratio = 1.0 self.xscale = xscale self.x, self.y = 0.0, 0.0 self.chart_width, self.chart_height = draw.extents(self.options, self.xscale, self.trace) self.hadj = None self.vadj = None self.hadj_changed_signal_id = None self.vadj_changed_signal_id = None def do_expose_event(self, event): cr = self.window.cairo_create() # set a clip region for the expose event cr.rectangle( event.area.x, event.area.y, event.area.width, event.area.height ) cr.clip() self.draw(cr, self.get_allocation()) return False def draw(self, cr, rect): cr.set_source_rgba(1.0, 1.0, 1.0, 1.0) cr.paint() cr.scale(self.zoom_ratio, self.zoom_ratio) cr.translate(-self.x, -self.y) draw.render(cr, self.options, self.xscale, self.trace) def position_changed(self): self.emit("position-changed", self.x, self.y) ZOOM_INCREMENT = 1.25 def zoom_image (self, zoom_ratio): self.zoom_ratio = zoom_ratio self._set_scroll_adjustments (self.hadj, self.vadj) self.queue_draw() def zoom_to_rect (self, rect): zoom_ratio = float(rect.width)/float(self.chart_width) self.zoom_image(zoom_ratio) self.x = 0 self.position_changed() def set_xscale(self, xscale): old_mid_x = self.x + self.hadj.page_size / 2 self.xscale = xscale self.chart_width, self.chart_height = draw.extents(self.options, self.xscale, self.trace) new_x = old_mid_x self.zoom_image (self.zoom_ratio) def on_expand(self, action): self.set_xscale (int(self.xscale * 1.5 + 0.5)) def on_contract(self, action): self.set_xscale (max(int(self.xscale / 1.5), 1)) def on_zoom_in(self, action): self.zoom_image(self.zoom_ratio * self.ZOOM_INCREMENT) def on_zoom_out(self, action): self.zoom_image(self.zoom_ratio / self.ZOOM_INCREMENT) def on_zoom_fit(self, action): self.zoom_to_rect(self.get_allocation()) def on_zoom_100(self, action): self.zoom_image(1.0) self.set_xscale(1.0) def show_toggled(self, button): self.options.app_options.show_all = button.get_property ('active') self.chart_width, self.chart_height = draw.extents(self.options, self.xscale, self.trace) self._set_scroll_adjustments(self.hadj, self.vadj) self.queue_draw() POS_INCREMENT = 100 def on_key_press_event(self, widget, event): if event.keyval == gtk.keysyms.Left: self.x -= self.POS_INCREMENT/self.zoom_ratio elif event.keyval == gtk.keysyms.Right: self.x += self.POS_INCREMENT/self.zoom_ratio elif event.keyval == gtk.keysyms.Up: self.y -= self.POS_INCREMENT/self.zoom_ratio elif event.keyval == gtk.keysyms.Down: self.y += self.POS_INCREMENT/self.zoom_ratio else: return False self.queue_draw() self.position_changed() return True def on_area_button_press(self, area, event): if event.button == 2 or event.button == 1: area.window.set_cursor(gtk.gdk.Cursor(gtk.gdk.FLEUR)) self.prevmousex = event.x self.prevmousey = event.y if event.type not in (gtk.gdk.BUTTON_PRESS, gtk.gdk.BUTTON_RELEASE): return False return False def on_area_button_release(self, area, event): if event.button == 2 or event.button == 1: area.window.set_cursor(gtk.gdk.Cursor(gtk.gdk.ARROW)) self.prevmousex = None self.prevmousey = None return True return False def on_area_scroll_event(self, area, event): if event.state & gtk.gdk.CONTROL_MASK: if event.direction == gtk.gdk.SCROLL_UP: self.zoom_image(self.zoom_ratio * self.ZOOM_INCREMENT) return True if event.direction == gtk.gdk.SCROLL_DOWN: self.zoom_image(self.zoom_ratio / self.ZOOM_INCREMENT) return True return False def on_area_motion_notify(self, area, event): state = event.state if state & gtk.gdk.BUTTON2_MASK or state & gtk.gdk.BUTTON1_MASK: x, y = int(event.x), int(event.y) # pan the image self.x += (self.prevmousex - x)/self.zoom_ratio self.y += (self.prevmousey - y)/self.zoom_ratio self.queue_draw() self.prevmousex = x self.prevmousey = y self.position_changed() return True def on_set_scroll_adjustments(self, area, hadj, vadj): self._set_scroll_adjustments (hadj, vadj) def on_allocation_size_changed(self, widget, allocation): self.hadj.page_size = allocation.width self.hadj.page_increment = allocation.width * 0.9 self.vadj.page_size = allocation.height self.vadj.page_increment = allocation.height * 0.9 def _set_adj_upper(self, adj, upper): changed = False value_changed = False if adj.upper != upper: adj.upper = upper changed = True max_value = max(0.0, upper - adj.page_size) if adj.value > max_value: adj.value = max_value value_changed = True if changed: adj.changed() if value_changed: adj.value_changed() def _set_scroll_adjustments(self, hadj, vadj): if hadj == None: hadj = gtk.Adjustment(0.0, 0.0, 0.0, 0.0, 0.0, 0.0) if vadj == None: vadj = gtk.Adjustment(0.0, 0.0, 0.0, 0.0, 0.0, 0.0) if self.hadj_changed_signal_id != None and \ self.hadj != None and hadj != self.hadj: self.hadj.disconnect (self.hadj_changed_signal_id) if self.vadj_changed_signal_id != None and \ self.vadj != None and vadj != self.vadj: self.vadj.disconnect (self.vadj_changed_signal_id) if hadj != None: self.hadj = hadj self._set_adj_upper (self.hadj, self.zoom_ratio * self.chart_width) self.hadj_changed_signal_id = self.hadj.connect('value-changed', self.on_adjustments_changed) if vadj != None: self.vadj = vadj self._set_adj_upper (self.vadj, self.zoom_ratio * self.chart_height) self.vadj_changed_signal_id = self.vadj.connect('value-changed', self.on_adjustments_changed) def on_adjustments_changed(self, adj): self.x = self.hadj.value / self.zoom_ratio self.y = self.vadj.value / self.zoom_ratio self.queue_draw() def on_position_changed(self, widget, x, y): self.hadj.value = x * self.zoom_ratio self.vadj.value = y * self.zoom_ratio PyBootchartWidget.set_set_scroll_adjustments_signal('set-scroll-adjustments') class PyBootchartShell(gtk.VBox): ui = ''' <ui> <toolbar name="ToolBar"> <toolitem action="Expand"/> <toolitem action="Contract"/> <separator/> <toolitem action="ZoomIn"/> <toolitem action="ZoomOut"/> <toolitem action="ZoomFit"/> <toolitem action="Zoom100"/> </toolbar> </ui> ''' def __init__(self, window, trace, options, xscale): gtk.VBox.__init__(self) self.widget = PyBootchartWidget(trace, options, xscale) # Create a UIManager instance uimanager = self.uimanager = gtk.UIManager() # Add the accelerator group to the toplevel window accelgroup = uimanager.get_accel_group() window.add_accel_group(accelgroup) # Create an ActionGroup actiongroup = gtk.ActionGroup('Actions') self.actiongroup = actiongroup # Create actions actiongroup.add_actions(( ('Expand', gtk.STOCK_ADD, None, None, None, self.widget.on_expand), ('Contract', gtk.STOCK_REMOVE, None, None, None, self.widget.on_contract), ('ZoomIn', gtk.STOCK_ZOOM_IN, None, None, None, self.widget.on_zoom_in), ('ZoomOut', gtk.STOCK_ZOOM_OUT, None, None, None, self.widget.on_zoom_out), ('ZoomFit', gtk.STOCK_ZOOM_FIT, 'Fit Width', None, None, self.widget.on_zoom_fit), ('Zoom100', gtk.STOCK_ZOOM_100, None, None, None, self.widget.on_zoom_100), )) # Add the actiongroup to the uimanager uimanager.insert_action_group(actiongroup, 0) # Add a UI description uimanager.add_ui_from_string(self.ui) # Scrolled window scrolled = gtk.ScrolledWindow() scrolled.add(self.widget) # toolbar / h-box hbox = gtk.HBox(False, 8) # Create a Toolbar toolbar = uimanager.get_widget('/ToolBar') hbox.pack_start(toolbar, True, True) if not options.kernel_only: # Misc. options button = gtk.CheckButton("Show more") button.connect ('toggled', self.widget.show_toggled) button.set_active(options.app_options.show_all) hbox.pack_start (button, False, True) self.pack_start(hbox, False) self.pack_start(scrolled) self.show_all() def grab_focus(self, window): window.set_focus(self.widget) class PyBootchartWindow(gtk.Window): def __init__(self, trace, app_options): gtk.Window.__init__(self) window = self window.set_title("Bootchart %s" % trace.filename) window.set_default_size(750, 550) tab_page = gtk.Notebook() tab_page.show() window.add(tab_page) full_opts = RenderOptions(app_options) full_tree = PyBootchartShell(window, trace, full_opts, 1.0) tab_page.append_page (full_tree, gtk.Label("Full tree")) if trace.kernel is not None and len (trace.kernel) > 2: kernel_opts = RenderOptions(app_options) kernel_opts.cumulative = False kernel_opts.charts = False kernel_opts.kernel_only = True kernel_tree = PyBootchartShell(window, trace, kernel_opts, 5.0) tab_page.append_page (kernel_tree, gtk.Label("Kernel boot")) full_tree.grab_focus(self) self.show() def show(trace, options): win = PyBootchartWindow(trace, options) win.connect('destroy', gtk.main_quit) gtk.main() ########NEW FILE######## __FILENAME__ = main main.py.in ########NEW FILE######## __FILENAME__ = parsing # This file is part of pybootchartgui. # pybootchartgui is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # pybootchartgui is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # You should have received a copy of the GNU General Public License # along with pybootchartgui. If not, see <http://www.gnu.org/licenses/>. from __future__ import with_statement import os import string import re import sys import tarfile from time import clock from collections import defaultdict from functools import reduce from .samples import * from .process_tree import ProcessTree if sys.version_info >= (3, 0): long = int # Parsing produces as its end result a 'Trace' class Trace: def __init__(self, writer, paths, options): self.processes = {} self.start = {} self.end = {} self.min = None self.max = None self.headers = None self.disk_stats = None self.ps_stats = None self.taskstats = None self.cpu_stats = None self.cmdline = None self.kernel = None self.kernel_tree = None self.filename = None self.parent_map = None self.mem_stats = None if len(paths): parse_paths (writer, self, paths) if not self.valid(): raise ParseError("empty state: '%s' does not contain a valid bootchart" % ", ".join(paths)) if options.full_time: self.min = min(self.start.keys()) self.max = max(self.end.keys()) return # Turn that parsed information into something more useful # link processes into a tree of pointers, calculate statistics self.compile(writer) # Crop the chart to the end of the first idle period after the given # process if options.crop_after: idle = self.crop (writer, options.crop_after) else: idle = None # Annotate other times as the first start point of given process lists self.times = [ idle ] if options.annotate: for procnames in options.annotate: names = [x[:15] for x in procnames.split(",")] for proc in self.ps_stats.process_map.values(): if proc.cmd in names: self.times.append(proc.start_time) break else: self.times.append(None) self.proc_tree = ProcessTree(writer, self.kernel, self.ps_stats, self.ps_stats.sample_period, self.headers.get("profile.process"), options.prune, idle, self.taskstats, self.parent_map is not None) if self.kernel is not None: self.kernel_tree = ProcessTree(writer, self.kernel, None, 0, self.headers.get("profile.process"), False, None, None, True) def valid(self): return len(self.processes) != 0 return self.headers != None and self.disk_stats != None and \ self.ps_stats != None and self.cpu_stats != None def add_process(self, process, start, end): self.processes[process] = [start, end] if start not in self.start: self.start[start] = [] if process not in self.start[start]: self.start[start].append(process) if end not in self.end: self.end[end] = [] if process not in self.end[end]: self.end[end].append(process) def compile(self, writer): def find_parent_id_for(pid): if pid is 0: return 0 ppid = self.parent_map.get(pid) if ppid: # many of these double forks are so short lived # that we have no samples, or process info for them # so climb the parent hierarcy to find one if int (ppid * 1000) not in self.ps_stats.process_map: # print "Pid '%d' short lived with no process" % ppid ppid = find_parent_id_for (ppid) # else: # print "Pid '%d' has an entry" % ppid else: # print "Pid '%d' missing from pid map" % pid return 0 return ppid # merge in the cmdline data if self.cmdline is not None: for proc in self.ps_stats.process_map.values(): rpid = int (proc.pid // 1000) if rpid in self.cmdline: cmd = self.cmdline[rpid] proc.exe = cmd['exe'] proc.args = cmd['args'] # else: # print "proc %d '%s' not in cmdline" % (rpid, proc.exe) # re-parent any stray orphans if we can if self.parent_map is not None: for process in self.ps_stats.process_map.values(): ppid = find_parent_id_for (int(process.pid // 1000)) if ppid: process.ppid = ppid * 1000 # stitch the tree together with pointers for process in self.ps_stats.process_map.values(): process.set_parent (self.ps_stats.process_map) # count on fingers variously for process in self.ps_stats.process_map.values(): process.calc_stats (self.ps_stats.sample_period) def crop(self, writer, crop_after): def is_idle_at(util, start, j): k = j + 1 while k < len(util) and util[k][0] < start + 300: k += 1 k = min(k, len(util)-1) if util[j][1] >= 0.25: return False avgload = sum(u[1] for u in util[j:k+1]) / (k-j+1) if avgload < 0.25: return True else: return False def is_idle(util, start): for j in range(0, len(util)): if util[j][0] < start: continue return is_idle_at(util, start, j) else: return False names = [x[:15] for x in crop_after.split(",")] for proc in self.ps_stats.process_map.values(): if proc.cmd in names or proc.exe in names: writer.info("selected proc '%s' from list (start %d)" % (proc.cmd, proc.start_time)) break if proc is None: writer.warn("no selected crop proc '%s' in list" % crop_after) cpu_util = [(sample.time, sample.user + sample.sys + sample.io) for sample in self.cpu_stats] disk_util = [(sample.time, sample.util) for sample in self.disk_stats] idle = None for i in range(0, len(cpu_util)): if cpu_util[i][0] < proc.start_time: continue if is_idle_at(cpu_util, cpu_util[i][0], i) \ and is_idle(disk_util, cpu_util[i][0]): idle = cpu_util[i][0] break if idle is None: writer.warn ("not idle after proc '%s'" % crop_after) return None crop_at = idle + 300 writer.info ("cropping at time %d" % crop_at) while len (self.cpu_stats) \ and self.cpu_stats[-1].time > crop_at: self.cpu_stats.pop() while len (self.disk_stats) \ and self.disk_stats[-1].time > crop_at: self.disk_stats.pop() self.ps_stats.end_time = crop_at cropped_map = {} for key, value in self.ps_stats.process_map.items(): if (value.start_time <= crop_at): cropped_map[key] = value for proc in cropped_map.values(): proc.duration = min (proc.duration, crop_at - proc.start_time) while len (proc.samples) \ and proc.samples[-1].time > crop_at: proc.samples.pop() self.ps_stats.process_map = cropped_map return idle class ParseError(Exception): """Represents errors during parse of the bootchart.""" def __init__(self, value): self.value = value def __str__(self): return self.value def _parse_headers(file): """Parses the headers of the bootchart.""" def parse(acc, line): (headers, last) = acc if '=' in line: last, value = map (lambda x: x.strip(), line.split('=', 1)) else: value = line.strip() headers[last] += value return headers, last return reduce(parse, file.read().decode('utf-8').split('\n'), (defaultdict(str),''))[0] def _parse_timed_blocks(file): """Parses (ie., splits) a file into so-called timed-blocks. A timed-block consists of a timestamp on a line by itself followed by zero or more lines of data for that point in time.""" def parse(block): lines = block.split('\n') if not lines: raise ParseError('expected a timed-block consisting a timestamp followed by data lines') try: return (int(lines[0]), lines[1:]) except ValueError: raise ParseError("expected a timed-block, but timestamp '%s' is not an integer" % lines[0]) blocks = file.read().decode('utf-8').split('\n\n') return [parse(block) for block in blocks if block.strip() and not block.endswith(' not running\n')] def _parse_proc_ps_log(writer, file): """ * See proc(5) for details. * * {pid, comm, state, ppid, pgrp, session, tty_nr, tpgid, flags, minflt, cminflt, majflt, cmajflt, utime, stime, * cutime, cstime, priority, nice, 0, itrealvalue, starttime, vsize, rss, rlim, startcode, endcode, startstack, * kstkesp, kstkeip} """ processMap = {} ltime = 0 timed_blocks = _parse_timed_blocks(file) for time, lines in timed_blocks: for line in lines: if not line: continue tokens = line.split(' ') if len(tokens) < 21: continue offset = [index for index, token in enumerate(tokens[1:]) if token[-1] == ')'][0] pid, cmd, state, ppid = int(tokens[0]), ' '.join(tokens[1:2+offset]), tokens[2+offset], int(tokens[3+offset]) userCpu, sysCpu, stime = int(tokens[13+offset]), int(tokens[14+offset]), int(tokens[21+offset]) # magic fixed point-ness ... pid *= 1000 ppid *= 1000 if pid in processMap: process = processMap[pid] process.cmd = cmd.strip('()') # why rename after latest name?? else: process = Process(writer, pid, cmd.strip('()'), ppid, min(time, stime)) processMap[pid] = process if process.last_user_cpu_time is not None and process.last_sys_cpu_time is not None and ltime is not None: userCpuLoad, sysCpuLoad = process.calc_load(userCpu, sysCpu, max(1, time - ltime)) cpuSample = CPUSample('null', userCpuLoad, sysCpuLoad, 0.0) process.samples.append(ProcessSample(time, state, cpuSample)) process.last_user_cpu_time = userCpu process.last_sys_cpu_time = sysCpu ltime = time if len (timed_blocks) < 2: return None startTime = timed_blocks[0][0] avgSampleLength = (ltime - startTime)/(len (timed_blocks) - 1) return ProcessStats (writer, processMap, len (timed_blocks), avgSampleLength, startTime, ltime) def _parse_taskstats_log(writer, file): """ * See bootchart-collector.c for details. * * { pid, ppid, comm, cpu_run_real_total, blkio_delay_total, swapin_delay_total } * """ processMap = {} pidRewrites = {} ltime = None timed_blocks = _parse_timed_blocks(file) for time, lines in timed_blocks: # we have no 'stime' from taskstats, so prep 'init' if ltime is None: process = Process(writer, 1, '[init]', 0, 0) processMap[1000] = process ltime = time # continue for line in lines: if not line: continue tokens = line.split(' ') if len(tokens) != 6: continue opid, ppid, cmd = int(tokens[0]), int(tokens[1]), tokens[2] cpu_ns, blkio_delay_ns, swapin_delay_ns = long(tokens[-3]), long(tokens[-2]), long(tokens[-1]), # make space for trees of pids opid *= 1000 ppid *= 1000 # when the process name changes, we re-write the pid. if opid in pidRewrites: pid = pidRewrites[opid] else: pid = opid cmd = cmd.strip('(').strip(')') if pid in processMap: process = processMap[pid] if process.cmd != cmd: pid += 1 pidRewrites[opid] = pid # print "process mutation ! '%s' vs '%s' pid %s -> pid %s\n" % (process.cmd, cmd, opid, pid) process = process.split (writer, pid, cmd, ppid, time) processMap[pid] = process else: process.cmd = cmd; else: process = Process(writer, pid, cmd, ppid, time) processMap[pid] = process delta_cpu_ns = (float) (cpu_ns - process.last_cpu_ns) delta_blkio_delay_ns = (float) (blkio_delay_ns - process.last_blkio_delay_ns) delta_swapin_delay_ns = (float) (swapin_delay_ns - process.last_swapin_delay_ns) # make up some state data ... if delta_cpu_ns > 0: state = "R" elif delta_blkio_delay_ns + delta_swapin_delay_ns > 0: state = "D" else: state = "S" # retain the ns timing information into a CPUSample - that tries # with the old-style to be a %age of CPU used in this time-slice. if delta_cpu_ns + delta_blkio_delay_ns + delta_swapin_delay_ns > 0: # print "proc %s cpu_ns %g delta_cpu %g" % (cmd, cpu_ns, delta_cpu_ns) cpuSample = CPUSample('null', delta_cpu_ns, 0.0, delta_blkio_delay_ns, delta_swapin_delay_ns) process.samples.append(ProcessSample(time, state, cpuSample)) process.last_cpu_ns = cpu_ns process.last_blkio_delay_ns = blkio_delay_ns process.last_swapin_delay_ns = swapin_delay_ns ltime = time if len (timed_blocks) < 2: return None startTime = timed_blocks[0][0] avgSampleLength = (ltime - startTime)/(len(timed_blocks)-1) return ProcessStats (writer, processMap, len (timed_blocks), avgSampleLength, startTime, ltime) def _parse_proc_stat_log(file): samples = [] ltimes = None for time, lines in _parse_timed_blocks(file): # skip emtpy lines if not lines: continue # CPU times {user, nice, system, idle, io_wait, irq, softirq} tokens = lines[0].split() times = [ int(token) for token in tokens[1:] ] if ltimes: user = float((times[0] + times[1]) - (ltimes[0] + ltimes[1])) system = float((times[2] + times[5] + times[6]) - (ltimes[2] + ltimes[5] + ltimes[6])) idle = float(times[3] - ltimes[3]) iowait = float(times[4] - ltimes[4]) aSum = max(user + system + idle + iowait, 1) samples.append( CPUSample(time, user/aSum, system/aSum, iowait/aSum) ) ltimes = times # skip the rest of statistics lines return samples def _parse_proc_disk_stat_log(file, numCpu): """ Parse file for disk stats, but only look at the whole device, eg. sda, not sda1, sda2 etc. The format of relevant lines should be: {major minor name rio rmerge rsect ruse wio wmerge wsect wuse running use aveq} """ disk_regex_re = re.compile ('^([hsv]d.|mtdblock\d|mmcblk\d|cciss/c\d+d\d+.*)$') # this gets called an awful lot. def is_relevant_line(linetokens): if len(linetokens) != 14: return False disk = linetokens[2] return disk_regex_re.match(disk) disk_stat_samples = [] for time, lines in _parse_timed_blocks(file): sample = DiskStatSample(time) relevant_tokens = [linetokens for linetokens in map (lambda x: x.split(),lines) if is_relevant_line(linetokens)] for tokens in relevant_tokens: disk, rsect, wsect, use = tokens[2], int(tokens[5]), int(tokens[9]), int(tokens[12]) sample.add_diskdata([rsect, wsect, use]) disk_stat_samples.append(sample) disk_stats = [] for sample1, sample2 in zip(disk_stat_samples[:-1], disk_stat_samples[1:]): interval = sample1.time - sample2.time if interval == 0: interval = 1 sums = [ a - b for a, b in zip(sample1.diskdata, sample2.diskdata) ] readTput = sums[0] / 2.0 * 100.0 / interval writeTput = sums[1] / 2.0 * 100.0 / interval util = float( sums[2] ) / 10 / interval / numCpu util = max(0.0, min(1.0, util)) disk_stats.append(DiskSample(sample2.time, readTput, writeTput, util)) return disk_stats def _parse_proc_meminfo_log(file): """ Parse file for global memory statistics. The format of relevant lines should be: ^key: value( unit)? """ used_values = ('MemTotal', 'MemFree', 'Buffers', 'Cached', 'SwapTotal', 'SwapFree',) mem_stats = [] meminfo_re = re.compile(r'([^ \t:]+):\s*(\d+).*') for time, lines in _parse_timed_blocks(file): sample = MemSample(time) for line in lines: match = meminfo_re.match(line) if not match: raise ParseError("Invalid meminfo line \"%s\"" % match.groups(0)) sample.add_value(match.group(1), int(match.group(2))) if sample.valid(): mem_stats.append(sample) return mem_stats # if we boot the kernel with: initcall_debug printk.time=1 we can # get all manner of interesting data from the dmesg output # We turn this into a pseudo-process tree: each event is # characterised by a # we don't try to detect a "kernel finished" state - since the kernel # continues to do interesting things after init is called. # # sample input: # [ 0.000000] ACPI: FACP 3f4fc000 000F4 (v04 INTEL Napa 00000001 MSFT 01000013) # ... # [ 0.039993] calling migration_init+0x0/0x6b @ 1 # [ 0.039993] initcall migration_init+0x0/0x6b returned 1 after 0 usecs def _parse_dmesg(writer, file): timestamp_re = re.compile ("^\[\s*(\d+\.\d+)\s*]\s+(.*)$") split_re = re.compile ("^(\S+)\s+([\S\+_-]+) (.*)$") processMap = {} idx = 0 inc = 1.0 / 1000000 kernel = Process(writer, idx, "k-boot", 0, 0.1) processMap['k-boot'] = kernel base_ts = False max_ts = 0 for line in file.read().decode('utf-8').split('\n'): t = timestamp_re.match (line) if t is None: # print "duff timestamp " + line continue time_ms = float (t.group(1)) * 1000 # looks like we may have a huge diff after the clock # has been set up. This could lead to huge graph: # so huge we will be killed by the OOM. # So instead of using the plain timestamp we will # use a delta to first one and skip the first one # for convenience if max_ts == 0 and not base_ts and time_ms > 1000: base_ts = time_ms continue max_ts = max(time_ms, max_ts) if base_ts: # print "fscked clock: used %f instead of %f" % (time_ms - base_ts, time_ms) time_ms -= base_ts m = split_re.match (t.group(2)) if m is None: continue # print "match: '%s'" % (m.group(1)) type = m.group(1) func = m.group(2) rest = m.group(3) if t.group(2).startswith ('Write protecting the') or \ t.group(2).startswith ('Freeing unused kernel memory'): kernel.duration = time_ms / 10 continue # print "foo: '%s' '%s' '%s'" % (type, func, rest) if type == "calling": ppid = kernel.pid p = re.match ("\@ (\d+)", rest) if p is not None: ppid = float (p.group(1)) // 1000 # print "match: '%s' ('%g') at '%s'" % (func, ppid, time_ms) name = func.split ('+', 1) [0] idx += inc processMap[func] = Process(writer, ppid + idx, name, ppid, time_ms / 10) elif type == "initcall": # print "finished: '%s' at '%s'" % (func, time_ms) if func in processMap: process = processMap[func] process.duration = (time_ms / 10) - process.start_time else: print("corrupted init call for %s" % (func)) elif type == "async_waiting" or type == "async_continuing": continue # ignore return processMap.values() # # Parse binary pacct accounting file output if we have one # cf. /usr/include/linux/acct.h # def _parse_pacct(writer, file): # read LE int32 def _read_le_int32(file): byts = file.read(4) return (ord(byts[0])) | (ord(byts[1]) << 8) | \ (ord(byts[2]) << 16) | (ord(byts[3]) << 24) parent_map = {} parent_map[0] = 0 while file.read(1) != "": # ignore flags ver = file.read(1) if ord(ver) < 3: print("Invalid version 0x%x" % (ord(ver))) return None file.seek (14, 1) # user, group etc. pid = _read_le_int32 (file) ppid = _read_le_int32 (file) # print "Parent of %d is %d" % (pid, ppid) parent_map[pid] = ppid file.seek (4 + 4 + 16, 1) # timings file.seek (16, 1) # acct_comm return parent_map def _parse_paternity_log(writer, file): parent_map = {} parent_map[0] = 0 for line in file.read().decode('utf-8').split('\n'): if not line: continue elems = line.split(' ') # <Child> <Parent> if len (elems) >= 2: # print "paternity of %d is %d" % (int(elems[0]), int(elems[1])) parent_map[int(elems[0])] = int(elems[1]) else: print("Odd paternity line '%s'" % (line)) return parent_map def _parse_cmdline_log(writer, file): cmdLines = {} for block in file.read().decode('utf-8').split('\n\n'): lines = block.split('\n') if len (lines) >= 3: # print "Lines '%s'" % (lines[0]) pid = int (lines[0]) values = {} values['exe'] = lines[1].lstrip(':') args = lines[2].lstrip(':').split('\0') args.pop() values['args'] = args cmdLines[pid] = values return cmdLines def get_num_cpus(headers): """Get the number of CPUs from the system.cpu header property. As the CPU utilization graphs are relative, the number of CPUs currently makes no difference.""" if headers is None: return 1 if headers.get("system.cpu.num"): return max (int (headers.get("system.cpu.num")), 1) cpu_model = headers.get("system.cpu") if cpu_model is None: return 1 mat = re.match(".*\\((\\d+)\\)", cpu_model) if mat is None: return 1 return max (int(mat.group(1)), 1) def _do_parse(writer, state, filename, file): writer.info("parsing '%s'" % filename) t1 = clock() paths = filename.split("/") task = paths[-1] pn = paths[-2] start = None end = None for line in file: if line.startswith("Started:"): start = int(float(line.split()[-1])) elif line.startswith("Ended:"): end = int(float(line.split()[-1])) if start and end: state.add_process(pn + ":" + task, start, end) t2 = clock() writer.info(" %s seconds" % str(t2-t1)) return state def parse_file(writer, state, filename): if state.filename is None: state.filename = filename basename = os.path.basename(filename) with open(filename, "rb") as file: return _do_parse(writer, state, filename, file) def parse_paths(writer, state, paths): for path in paths: if state.filename is None: state.filename = path root, extension = os.path.splitext(path) if not(os.path.exists(path)): writer.warn("warning: path '%s' does not exist, ignoring." % path) continue #state.filename = path if os.path.isdir(path): files = sorted([os.path.join(path, f) for f in os.listdir(path)]) state = parse_paths(writer, state, files) elif extension in [".tar", ".tgz", ".gz"]: if extension == ".gz": root, extension = os.path.splitext(root) if extension != ".tar": writer.warn("warning: can only handle zipped tar files, not zipped '%s'-files; ignoring" % extension) continue tf = None try: writer.status("parsing '%s'" % path) tf = tarfile.open(path, 'r:*') for name in tf.getnames(): state = _do_parse(writer, state, name, tf.extractfile(name)) except tarfile.ReadError as error: raise ParseError("error: could not read tarfile '%s': %s." % (path, error)) finally: if tf != None: tf.close() else: state = parse_file(writer, state, path) return state def split_res(res, options): """ Split the res into n pieces """ res_list = [] if options.num > 1: s_list = sorted(res.start.keys()) frag_size = len(s_list) / float(options.num) # Need the top value if frag_size > int(frag_size): frag_size = int(frag_size + 1) else: frag_size = int(frag_size) start = 0 end = frag_size while start < end: state = Trace(None, [], None) if options.full_time: state.min = min(res.start.keys()) state.max = max(res.end.keys()) for i in range(start, end): # Add this line for reference #state.add_process(pn + ":" + task, start, end) for p in res.start[s_list[i]]: state.add_process(p, s_list[i], res.processes[p][1]) start = end end = end + frag_size if end > len(s_list): end = len(s_list) res_list.append(state) else: res_list.append(res) return res_list ########NEW FILE######## __FILENAME__ = process_tree # This file is part of pybootchartgui. # pybootchartgui is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # pybootchartgui is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # You should have received a copy of the GNU General Public License # along with pybootchartgui. If not, see <http://www.gnu.org/licenses/>. class ProcessTree: """ProcessTree encapsulates a process tree. The tree is built from log files retrieved during the boot process. When building the process tree, it is pruned and merged in order to be able to visualize it in a comprehensible manner. The following pruning techniques are used: * idle processes that keep running during the last process sample (which is a heuristic for a background processes) are removed, * short-lived processes (i.e. processes that only live for the duration of two samples or less) are removed, * the processes used by the boot logger are removed, * exploders (i.e. processes that are known to spawn huge meaningless process subtrees) have their subtrees merged together, * siblings (i.e. processes with the same command line living concurrently -- thread heuristic) are merged together, * process runs (unary trees with processes sharing the command line) are merged together. """ LOGGER_PROC = 'bootchart-colle' EXPLODER_PROCESSES = set(['hwup']) def __init__(self, writer, kernel, psstats, sample_period, monitoredApp, prune, idle, taskstats, accurate_parentage, for_testing = False): self.writer = writer self.process_tree = [] self.taskstats = taskstats if psstats is None: process_list = kernel elif kernel is None: process_list = psstats.process_map.values() else: process_list = list(kernel) + list(psstats.process_map.values()) self.process_list = sorted(process_list, key = lambda p: p.pid) self.sample_period = sample_period self.build() if not accurate_parentage: self.update_ppids_for_daemons(self.process_list) self.start_time = self.get_start_time(self.process_tree) self.end_time = self.get_end_time(self.process_tree) self.duration = self.end_time - self.start_time self.idle = idle if for_testing: return removed = self.merge_logger(self.process_tree, self.LOGGER_PROC, monitoredApp, False) writer.status("merged %i logger processes" % removed) if prune: p_processes = self.prune(self.process_tree, None) p_exploders = self.merge_exploders(self.process_tree, self.EXPLODER_PROCESSES) p_threads = self.merge_siblings(self.process_tree) p_runs = self.merge_runs(self.process_tree) writer.status("pruned %i process, %i exploders, %i threads, and %i runs" % (p_processes, p_exploders, p_threads, p_runs)) self.sort(self.process_tree) self.start_time = self.get_start_time(self.process_tree) self.end_time = self.get_end_time(self.process_tree) self.duration = self.end_time - self.start_time self.num_proc = self.num_nodes(self.process_tree) def build(self): """Build the process tree from the list of top samples.""" self.process_tree = [] for proc in self.process_list: if not proc.parent: self.process_tree.append(proc) else: proc.parent.child_list.append(proc) def sort(self, process_subtree): """Sort process tree.""" for p in process_subtree: p.child_list.sort(key = lambda p: p.pid) self.sort(p.child_list) def num_nodes(self, process_list): "Counts the number of nodes in the specified process tree.""" nodes = 0 for proc in process_list: nodes = nodes + self.num_nodes(proc.child_list) return nodes + len(process_list) def get_start_time(self, process_subtree): """Returns the start time of the process subtree. This is the start time of the earliest process. """ if not process_subtree: return 100000000 return min( [min(proc.start_time, self.get_start_time(proc.child_list)) for proc in process_subtree] ) def get_end_time(self, process_subtree): """Returns the end time of the process subtree. This is the end time of the last collected sample. """ if not process_subtree: return -100000000 return max( [max(proc.start_time + proc.duration, self.get_end_time(proc.child_list)) for proc in process_subtree] ) def get_max_pid(self, process_subtree): """Returns the max PID found in the process tree.""" if not process_subtree: return -100000000 return max( [max(proc.pid, self.get_max_pid(proc.child_list)) for proc in process_subtree] ) def update_ppids_for_daemons(self, process_list): """Fedora hack: when loading the system services from rc, runuser(1) is used. This sets the PPID of all daemons to 1, skewing the process tree. Try to detect this and set the PPID of these processes the PID of rc. """ rcstartpid = -1 rcendpid = -1 rcproc = None for p in process_list: if p.cmd == "rc" and p.ppid // 1000 == 1: rcproc = p rcstartpid = p.pid rcendpid = self.get_max_pid(p.child_list) if rcstartpid != -1 and rcendpid != -1: for p in process_list: if p.pid > rcstartpid and p.pid < rcendpid and p.ppid // 1000 == 1: p.ppid = rcstartpid p.parent = rcproc for p in process_list: p.child_list = [] self.build() def prune(self, process_subtree, parent): """Prunes the process tree by removing idle processes and processes that only live for the duration of a single top sample. Sibling processes with the same command line (i.e. threads) are merged together. This filters out sleepy background processes, short-lived processes and bootcharts' analysis tools. """ def is_idle_background_process_without_children(p): process_end = p.start_time + p.duration return not p.active and \ process_end >= self.start_time + self.duration and \ p.start_time > self.start_time and \ p.duration > 0.9 * self.duration and \ self.num_nodes(p.child_list) == 0 num_removed = 0 idx = 0 while idx < len(process_subtree): p = process_subtree[idx] if parent != None or len(p.child_list) == 0: prune = False if is_idle_background_process_without_children(p): prune = True elif p.duration <= 2 * self.sample_period: # short-lived process prune = True if prune: process_subtree.pop(idx) for c in p.child_list: process_subtree.insert(idx, c) num_removed += 1 continue else: num_removed += self.prune(p.child_list, p) else: num_removed += self.prune(p.child_list, p) idx += 1 return num_removed def merge_logger(self, process_subtree, logger_proc, monitored_app, app_tree): """Merges the logger's process subtree. The logger will typically spawn lots of sleep and cat processes, thus polluting the process tree. """ num_removed = 0 for p in process_subtree: is_app_tree = app_tree if logger_proc == p.cmd and not app_tree: is_app_tree = True num_removed += self.merge_logger(p.child_list, logger_proc, monitored_app, is_app_tree) # don't remove the logger itself continue if app_tree and monitored_app != None and monitored_app == p.cmd: is_app_tree = False if is_app_tree: for child in p.child_list: self.merge_processes(p, child) num_removed += 1 p.child_list = [] else: num_removed += self.merge_logger(p.child_list, logger_proc, monitored_app, is_app_tree) return num_removed def merge_exploders(self, process_subtree, processes): """Merges specific process subtrees (used for processes which usually spawn huge meaningless process trees). """ num_removed = 0 for p in process_subtree: if processes in processes and len(p.child_list) > 0: subtreemap = self.getProcessMap(p.child_list) for child in subtreemap.values(): self.merge_processes(p, child) num_removed += len(subtreemap) p.child_list = [] p.cmd += " (+)" else: num_removed += self.merge_exploders(p.child_list, processes) return num_removed def merge_siblings(self, process_subtree): """Merges thread processes. Sibling processes with the same command line are merged together. """ num_removed = 0 idx = 0 while idx < len(process_subtree)-1: p = process_subtree[idx] nextp = process_subtree[idx+1] if nextp.cmd == p.cmd: process_subtree.pop(idx+1) idx -= 1 num_removed += 1 p.child_list.extend(nextp.child_list) self.merge_processes(p, nextp) num_removed += self.merge_siblings(p.child_list) idx += 1 if len(process_subtree) > 0: p = process_subtree[-1] num_removed += self.merge_siblings(p.child_list) return num_removed def merge_runs(self, process_subtree): """Merges process runs. Single child processes which share the same command line with the parent are merged. """ num_removed = 0 idx = 0 while idx < len(process_subtree): p = process_subtree[idx] if len(p.child_list) == 1 and p.child_list[0].cmd == p.cmd: child = p.child_list[0] p.child_list = list(child.child_list) self.merge_processes(p, child) num_removed += 1 continue num_removed += self.merge_runs(p.child_list) idx += 1 return num_removed def merge_processes(self, p1, p2): """Merges two process' samples.""" p1.samples.extend(p2.samples) p1.samples.sort( key = lambda p: p.time ) p1time = p1.start_time p2time = p2.start_time p1.start_time = min(p1time, p2time) pendtime = max(p1time + p1.duration, p2time + p2.duration) p1.duration = pendtime - p1.start_time ########NEW FILE######## __FILENAME__ = samples # This file is part of pybootchartgui. # pybootchartgui is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # pybootchartgui is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # You should have received a copy of the GNU General Public License # along with pybootchartgui. If not, see <http://www.gnu.org/licenses/>. class DiskStatSample: def __init__(self, time): self.time = time self.diskdata = [0, 0, 0] def add_diskdata(self, new_diskdata): self.diskdata = [ a + b for a, b in zip(self.diskdata, new_diskdata) ] class CPUSample: def __init__(self, time, user, sys, io = 0.0, swap = 0.0): self.time = time self.user = user self.sys = sys self.io = io self.swap = swap @property def cpu(self): return self.user + self.sys def __str__(self): return str(self.time) + "\t" + str(self.user) + "\t" + \ str(self.sys) + "\t" + str(self.io) + "\t" + str (self.swap) class MemSample: used_values = ('MemTotal', 'MemFree', 'Buffers', 'Cached', 'SwapTotal', 'SwapFree',) def __init__(self, time): self.time = time self.records = {} def add_value(self, name, value): if name in MemSample.used_values: self.records[name] = value def valid(self): keys = self.records.keys() # discard incomplete samples return [v for v in MemSample.used_values if v not in keys] == [] class ProcessSample: def __init__(self, time, state, cpu_sample): self.time = time self.state = state self.cpu_sample = cpu_sample def __str__(self): return str(self.time) + "\t" + str(self.state) + "\t" + str(self.cpu_sample) class ProcessStats: def __init__(self, writer, process_map, sample_count, sample_period, start_time, end_time): self.process_map = process_map self.sample_count = sample_count self.sample_period = sample_period self.start_time = start_time self.end_time = end_time writer.info ("%d samples, avg. sample length %f" % (self.sample_count, self.sample_period)) writer.info ("process list size: %d" % len (self.process_map.values())) class Process: def __init__(self, writer, pid, cmd, ppid, start_time): self.writer = writer self.pid = pid self.cmd = cmd self.exe = cmd self.args = [] self.ppid = ppid self.start_time = start_time self.duration = 0 self.samples = [] self.parent = None self.child_list = [] self.active = None self.last_user_cpu_time = None self.last_sys_cpu_time = None self.last_cpu_ns = 0 self.last_blkio_delay_ns = 0 self.last_swapin_delay_ns = 0 # split this process' run - triggered by a name change def split(self, writer, pid, cmd, ppid, start_time): split = Process (writer, pid, cmd, ppid, start_time) split.last_cpu_ns = self.last_cpu_ns split.last_blkio_delay_ns = self.last_blkio_delay_ns split.last_swapin_delay_ns = self.last_swapin_delay_ns return split def __str__(self): return " ".join([str(self.pid), self.cmd, str(self.ppid), '[ ' + str(len(self.samples)) + ' samples ]' ]) def calc_stats(self, samplePeriod): if self.samples: firstSample = self.samples[0] lastSample = self.samples[-1] self.start_time = min(firstSample.time, self.start_time) self.duration = lastSample.time - self.start_time + samplePeriod activeCount = sum( [1 for sample in self.samples if sample.cpu_sample and sample.cpu_sample.sys + sample.cpu_sample.user + sample.cpu_sample.io > 0.0] ) activeCount = activeCount + sum( [1 for sample in self.samples if sample.state == 'D'] ) self.active = (activeCount>2) def calc_load(self, userCpu, sysCpu, interval): userCpuLoad = float(userCpu - self.last_user_cpu_time) / interval sysCpuLoad = float(sysCpu - self.last_sys_cpu_time) / interval cpuLoad = userCpuLoad + sysCpuLoad # normalize if cpuLoad > 1.0: userCpuLoad = userCpuLoad / cpuLoad sysCpuLoad = sysCpuLoad / cpuLoad return (userCpuLoad, sysCpuLoad) def set_parent(self, processMap): if self.ppid != None: self.parent = processMap.get (self.ppid) if self.parent == None and self.pid // 1000 > 1 and \ not (self.ppid == 2000 or self.pid == 2000): # kernel threads: ppid=2 self.writer.warn("Missing CONFIG_PROC_EVENTS: no parent for pid '%i' ('%s') with ppid '%i'" \ % (self.pid,self.cmd,self.ppid)) def get_end_time(self): return self.start_time + self.duration class DiskSample: def __init__(self, time, read, write, util): self.time = time self.read = read self.write = write self.util = util self.tput = read + write def __str__(self): return "\t".join([str(self.time), str(self.read), str(self.write), str(self.util)]) ########NEW FILE######## __FILENAME__ = parser_test import sys, os, re, struct, operator, math from collections import defaultdict import unittest sys.path.insert(0, os.getcwd()) import pybootchartgui.parsing as parsing import pybootchartgui.main as main debug = False def floatEq(f1, f2): return math.fabs(f1-f2) < 0.00001 bootchart_dir = os.path.join(os.path.dirname(sys.argv[0]), '../../examples/1/') parser = main._mk_options_parser() options, args = parser.parse_args(['--q', bootchart_dir]) writer = main._mk_writer(options) class TestBCParser(unittest.TestCase): def setUp(self): self.name = "My first unittest" self.rootdir = bootchart_dir def mk_fname(self,f): return os.path.join(self.rootdir, f) def testParseHeader(self): trace = parsing.Trace(writer, args, options) state = parsing.parse_file(writer, trace, self.mk_fname('header')) self.assertEqual(6, len(state.headers)) self.assertEqual(2, parsing.get_num_cpus(state.headers)) def test_parseTimedBlocks(self): trace = parsing.Trace(writer, args, options) state = parsing.parse_file(writer, trace, self.mk_fname('proc_diskstats.log')) self.assertEqual(141, len(state.disk_stats)) def testParseProcPsLog(self): trace = parsing.Trace(writer, args, options) state = parsing.parse_file(writer, trace, self.mk_fname('proc_ps.log')) samples = state.ps_stats processes = samples.process_map sorted_processes = [processes[k] for k in sorted(processes.keys())] ps_data = open(self.mk_fname('extract2.proc_ps.log')) for index, line in enumerate(ps_data): tokens = line.split(); process = sorted_processes[index] if debug: print(tokens[0:4]) print(process.pid / 1000, process.cmd, process.ppid, len(process.samples)) print('-------------------') self.assertEqual(tokens[0], str(process.pid // 1000)) self.assertEqual(tokens[1], str(process.cmd)) self.assertEqual(tokens[2], str(process.ppid // 1000)) self.assertEqual(tokens[3], str(len(process.samples))) ps_data.close() def testparseProcDiskStatLog(self): trace = parsing.Trace(writer, args, options) state_with_headers = parsing.parse_file(writer, trace, self.mk_fname('header')) state_with_headers.headers['system.cpu'] = 'xxx (2)' samples = parsing.parse_file(writer, state_with_headers, self.mk_fname('proc_diskstats.log')).disk_stats self.assertEqual(141, len(samples)) diskstats_data = open(self.mk_fname('extract.proc_diskstats.log')) for index, line in enumerate(diskstats_data): tokens = line.split('\t') sample = samples[index] if debug: print(line.rstrip()) print(sample) print('-------------------') self.assertEqual(tokens[0], str(sample.time)) self.assert_(floatEq(float(tokens[1]), sample.read)) self.assert_(floatEq(float(tokens[2]), sample.write)) self.assert_(floatEq(float(tokens[3]), sample.util)) diskstats_data.close() def testparseProcStatLog(self): trace = parsing.Trace(writer, args, options) samples = parsing.parse_file(writer, trace, self.mk_fname('proc_stat.log')).cpu_stats self.assertEqual(141, len(samples)) stat_data = open(self.mk_fname('extract.proc_stat.log')) for index, line in enumerate(stat_data): tokens = line.split('\t') sample = samples[index] if debug: print(line.rstrip()) print(sample) print('-------------------') self.assert_(floatEq(float(tokens[0]), sample.time)) self.assert_(floatEq(float(tokens[1]), sample.user)) self.assert_(floatEq(float(tokens[2]), sample.sys)) self.assert_(floatEq(float(tokens[3]), sample.io)) stat_data.close() if __name__ == '__main__': unittest.main() ########NEW FILE######## __FILENAME__ = process_tree_test import sys import os import unittest sys.path.insert(0, os.getcwd()) import pybootchartgui.parsing as parsing import pybootchartgui.process_tree as process_tree import pybootchartgui.main as main if sys.version_info >= (3, 0): long = int class TestProcessTree(unittest.TestCase): def setUp(self): self.name = "Process tree unittest" self.rootdir = os.path.join(os.path.dirname(sys.argv[0]), '../../examples/1/') parser = main._mk_options_parser() options, args = parser.parse_args(['--q', self.rootdir]) writer = main._mk_writer(options) trace = parsing.Trace(writer, args, options) parsing.parse_file(writer, trace, self.mk_fname('proc_ps.log')) trace.compile(writer) self.processtree = process_tree.ProcessTree(writer, None, trace.ps_stats, \ trace.ps_stats.sample_period, None, options.prune, None, None, False, for_testing = True) def mk_fname(self,f): return os.path.join(self.rootdir, f) def flatten(self, process_tree): flattened = [] for p in process_tree: flattened.append(p) flattened.extend(self.flatten(p.child_list)) return flattened def checkAgainstJavaExtract(self, filename, process_tree): test_data = open(filename) for expected, actual in zip(test_data, self.flatten(process_tree)): tokens = expected.split('\t') self.assertEqual(int(tokens[0]), actual.pid // 1000) self.assertEqual(tokens[1], actual.cmd) self.assertEqual(long(tokens[2]), 10 * actual.start_time) self.assert_(long(tokens[3]) - 10 * actual.duration < 5, "duration") self.assertEqual(int(tokens[4]), len(actual.child_list)) self.assertEqual(int(tokens[5]), len(actual.samples)) test_data.close() def testBuild(self): process_tree = self.processtree.process_tree self.checkAgainstJavaExtract(self.mk_fname('extract.processtree.1.log'), process_tree) def testMergeLogger(self): self.processtree.merge_logger(self.processtree.process_tree, 'bootchartd', None, False) process_tree = self.processtree.process_tree self.checkAgainstJavaExtract(self.mk_fname('extract.processtree.2.log'), process_tree) def testPrune(self): self.processtree.merge_logger(self.processtree.process_tree, 'bootchartd', None, False) self.processtree.prune(self.processtree.process_tree, None) process_tree = self.processtree.process_tree self.checkAgainstJavaExtract(self.mk_fname('extract.processtree.3b.log'), process_tree) def testMergeExploders(self): self.processtree.merge_logger(self.processtree.process_tree, 'bootchartd', None, False) self.processtree.prune(self.processtree.process_tree, None) self.processtree.merge_exploders(self.processtree.process_tree, set(['hwup'])) process_tree = self.processtree.process_tree self.checkAgainstJavaExtract(self.mk_fname('extract.processtree.3c.log'), process_tree) def testMergeSiblings(self): self.processtree.merge_logger(self.processtree.process_tree, 'bootchartd', None, False) self.processtree.prune(self.processtree.process_tree, None) self.processtree.merge_exploders(self.processtree.process_tree, set(['hwup'])) self.processtree.merge_siblings(self.processtree.process_tree) process_tree = self.processtree.process_tree self.checkAgainstJavaExtract(self.mk_fname('extract.processtree.3d.log'), process_tree) def testMergeRuns(self): self.processtree.merge_logger(self.processtree.process_tree, 'bootchartd', None, False) self.processtree.prune(self.processtree.process_tree, None) self.processtree.merge_exploders(self.processtree.process_tree, set(['hwup'])) self.processtree.merge_siblings(self.processtree.process_tree) self.processtree.merge_runs(self.processtree.process_tree) process_tree = self.processtree.process_tree self.checkAgainstJavaExtract(self.mk_fname('extract.processtree.3e.log'), process_tree) if __name__ == '__main__': unittest.main() ########NEW FILE######## __FILENAME__ = pybootchartgui #!/usr/bin/python # # This file is part of pybootchartgui. # pybootchartgui is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # pybootchartgui is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # You should have received a copy of the GNU General Public License # along with pybootchartgui. If not, see <http://www.gnu.org/licenses/>. import sys from pybootchartgui.main import main if __name__ == '__main__': sys.exit(main()) ########NEW FILE######## __FILENAME__ = relocate_sdk #!/usr/bin/env python # # Copyright (c) 2012 Intel Corporation # # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License version 2 as # published by the Free Software Foundation. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # See the GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program; if not, write to the Free Software # Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA # # DESCRIPTION # This script is called by the SDK installer script. It replaces the dynamic # loader path in all binaries and also fixes the SYSDIR paths/lengths and the # location of ld.so.cache in the dynamic loader binary # # AUTHORS # Laurentiu Palcu <laurentiu.palcu@intel.com> # import struct import sys import stat import os import re import errno if sys.version < '3': def b(x): return x else: def b(x): return x.encode(sys.getfilesystemencoding()) old_prefix = re.compile(b("##DEFAULT_INSTALL_DIR##")) def get_arch(): f.seek(0) e_ident =f.read(16) ei_mag0,ei_mag1_3,ei_class = struct.unpack("<B3sB11x", e_ident) if (ei_mag0 != 0x7f and ei_mag1_3 != "ELF") or ei_class == 0: return 0 if ei_class == 1: return 32 elif ei_class == 2: return 64 def parse_elf_header(): global e_type, e_machine, e_version, e_entry, e_phoff, e_shoff, e_flags,\ e_ehsize, e_phentsize, e_phnum, e_shentsize, e_shnum, e_shstrndx f.seek(0) elf_header = f.read(64) if arch == 32: # 32bit hdr_fmt = "<HHILLLIHHHHHH" hdr_size = 52 else: # 64bit hdr_fmt = "<HHIQQQIHHHHHH" hdr_size = 64 e_type, e_machine, e_version, e_entry, e_phoff, e_shoff, e_flags,\ e_ehsize, e_phentsize, e_phnum, e_shentsize, e_shnum, e_shstrndx =\ struct.unpack(hdr_fmt, elf_header[16:hdr_size]) def change_interpreter(elf_file_name): if arch == 32: ph_fmt = "<IIIIIIII" else: ph_fmt = "<IIQQQQQQ" """ look for PT_INTERP section """ for i in range(0,e_phnum): f.seek(e_phoff + i * e_phentsize) ph_hdr = f.read(e_phentsize) if arch == 32: # 32bit p_type, p_offset, p_vaddr, p_paddr, p_filesz,\ p_memsz, p_flags, p_align = struct.unpack(ph_fmt, ph_hdr) else: # 64bit p_type, p_flags, p_offset, p_vaddr, p_paddr, \ p_filesz, p_memsz, p_align = struct.unpack(ph_fmt, ph_hdr) """ change interpreter """ if p_type == 3: # PT_INTERP section f.seek(p_offset) # External SDKs with mixed pre-compiled binaries should not get # relocated so look for some variant of /lib fname = f.read(11) if fname.startswith(b("/lib/")) or fname.startswith(b("/lib64/")) or \ fname.startswith(b("/lib32/")) or fname.startswith(b("/usr/lib32/")) or \ fname.startswith(b("/usr/lib32/")) or fname.startswith(b("/usr/lib64/")): break if (len(new_dl_path) >= p_filesz): print("ERROR: could not relocate %s, interp size = %i and %i is needed." \ % (elf_file_name, p_memsz, len(new_dl_path) + 1)) break dl_path = new_dl_path + b("\0") * (p_filesz - len(new_dl_path)) f.seek(p_offset) f.write(dl_path) break def change_dl_sysdirs(): if arch == 32: sh_fmt = "<IIIIIIIIII" else: sh_fmt = "<IIQQQQIIQQ" """ read section string table """ f.seek(e_shoff + e_shstrndx * e_shentsize) sh_hdr = f.read(e_shentsize) if arch == 32: sh_offset, sh_size = struct.unpack("<16xII16x", sh_hdr) else: sh_offset, sh_size = struct.unpack("<24xQQ24x", sh_hdr) f.seek(sh_offset) sh_strtab = f.read(sh_size) sysdirs = sysdirs_len = "" """ change ld.so.cache path and default libs path for dynamic loader """ for i in range(0,e_shnum): f.seek(e_shoff + i * e_shentsize) sh_hdr = f.read(e_shentsize) sh_name, sh_type, sh_flags, sh_addr, sh_offset, sh_size, sh_link,\ sh_info, sh_addralign, sh_entsize = struct.unpack(sh_fmt, sh_hdr) name = sh_strtab[sh_name:sh_strtab.find(b("\0"), sh_name)] """ look only into SHT_PROGBITS sections """ if sh_type == 1: f.seek(sh_offset) """ default library paths cannot be changed on the fly because """ """ the string lengths have to be changed too. """ if name == b(".sysdirs"): sysdirs = f.read(sh_size) sysdirs_off = sh_offset sysdirs_sect_size = sh_size elif name == b(".sysdirslen"): sysdirslen = f.read(sh_size) sysdirslen_off = sh_offset elif name == b(".ldsocache"): ldsocache_path = f.read(sh_size) new_ldsocache_path = old_prefix.sub(new_prefix, ldsocache_path) # pad with zeros new_ldsocache_path += b("\0") * (sh_size - len(new_ldsocache_path)) # write it back f.seek(sh_offset) f.write(new_ldsocache_path) if sysdirs != "" and sysdirslen != "": paths = sysdirs.split(b("\0")) sysdirs = b("") sysdirslen = b("") for path in paths: """ exit the loop when we encounter first empty string """ if path == b(""): break new_path = old_prefix.sub(new_prefix, path) sysdirs += new_path + b("\0") if arch == 32: sysdirslen += struct.pack("<L", len(new_path)) else: sysdirslen += struct.pack("<Q", len(new_path)) """ pad with zeros """ sysdirs += b("\0") * (sysdirs_sect_size - len(sysdirs)) """ write the sections back """ f.seek(sysdirs_off) f.write(sysdirs) f.seek(sysdirslen_off) f.write(sysdirslen) # MAIN if len(sys.argv) < 4: sys.exit(-1) # In python > 3, strings may also contain Unicode characters. So, convert # them to bytes if sys.version_info < (3,): new_prefix = sys.argv[1] new_dl_path = sys.argv[2] else: new_prefix = sys.argv[1].encode() new_dl_path = sys.argv[2].encode() executables_list = sys.argv[3:] for e in executables_list: perms = os.stat(e)[stat.ST_MODE] if os.access(e, os.W_OK|os.R_OK): perms = None else: os.chmod(e, perms|stat.S_IRWXU) try: f = open(e, "r+b") except IOError: exctype, ioex = sys.exc_info()[:2] if ioex.errno == errno.ETXTBSY: print("Could not open %s. File used by another process.\nPlease "\ "make sure you exit all processes that might use any SDK "\ "binaries." % e) else: print("Could not open %s: %s(%d)" % (e, ioex.strerror, ioex.errno)) sys.exit(-1) # Save old size and do a size check at the end. Just a safety measure. old_size = os.path.getsize(e) arch = get_arch() if arch: parse_elf_header() change_interpreter(e) change_dl_sysdirs() """ change permissions back """ if perms: os.chmod(e, perms) f.close() if old_size != os.path.getsize(e): print("New file size for %s is different. Looks like a relocation error!", e) sys.exit(-1) ########NEW FILE######## __FILENAME__ = sysroot-relativelinks #!/usr/bin/env python import sys import os # Take a sysroot directory and turn all the abolute symlinks and turn them into # relative ones such that the sysroot is usable within another system. if len(sys.argv) != 2: print("Usage is " + sys.argv[0] + "<directory>") sys.exit(1) topdir = sys.argv[1] topdir = os.path.abspath(topdir) def handlelink(filep, subdir): link = os.readlink(filep) if link[0] != "/": return if link.startswith(topdir): return #print("Replacing %s with %s for %s" % (link, topdir+link, filep)) print("Replacing %s with %s for %s" % (link, os.path.relpath(topdir+link, subdir), filep)) os.unlink(filep) os.symlink(os.path.relpath(topdir+link, subdir), filep) for subdir, dirs, files in os.walk(topdir): for f in files: filep = os.path.join(subdir, f) if os.path.islink(filep): #print("Considering %s" % filep) handlelink(filep, subdir) ########NEW FILE######## __FILENAME__ = dirsize #!/usr/bin/env python # # Copyright (c) 2011, Intel Corporation. # All rights reserved. # # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation; either version 2 of the License, or # (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program; if not, write to the Free Software # Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. # # # Display details of the root filesystem size, broken up by directory. # Allows for limiting by size to focus on the larger files. # # Author: Darren Hart <dvhart@linux.intel.com> # import os import sys import stat class Record: def create(path): r = Record(path) s = os.lstat(path) if stat.S_ISDIR(s.st_mode): for p in os.listdir(path): pathname = path + "/" + p ss = os.lstat(pathname) if not stat.S_ISLNK(ss.st_mode): r.records.append(Record.create(pathname)) r.size += r.records[-1].size r.records.sort(reverse=True) else: r.size = os.lstat(path).st_size return r create = staticmethod(create) def __init__(self, path): self.path = path self.size = 0 self.records = [] def __cmp__(this, that): if that is None: return 1 if not isinstance(that, Record): raise TypeError if len(this.records) > 0 and len(that.records) == 0: return -1 if len(this.records) == 0 and len(that.records) > 0: return 1 if this.size < that.size: return -1 if this.size > that.size: return 1 return 0 def show(self, minsize): total = 0 if self.size <= minsize: return 0 print "%10d %s" % (self.size, self.path) for r in self.records: total += r.show(minsize) if len(self.records) == 0: total = self.size return total def main(): minsize = 0 if len(sys.argv) == 2: minsize = int(sys.argv[1]) rootfs = Record.create(".") total = rootfs.show(minsize) print "Displayed %d/%d bytes (%.2f%%)" % \ (total, rootfs.size, 100 * float(total) / rootfs.size) if __name__ == "__main__": main() ########NEW FILE######## __FILENAME__ = ksize #!/usr/bin/env python # # Copyright (c) 2011, Intel Corporation. # All rights reserved. # # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation; either version 2 of the License, or # (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program; if not, write to the Free Software # Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. # # # Display details of the kernel build size, broken up by built-in.o. Sort # the objects by size. Run from the top level kernel build directory. # # Author: Darren Hart <dvhart@linux.intel.com> # import sys import getopt import os from subprocess import * from string import join def usage(): prog = os.path.basename(sys.argv[0]) print 'Usage: %s [OPTION]...' % (prog) print ' -d, display an additional level of drivers detail' print ' -h, --help display this help and exit' print '' print 'Run %s from the top-level Linux kernel build directory.' % (prog) class Sizes: def __init__(self, glob): self.title = glob p = Popen("size -t " + glob, shell=True, stdout=PIPE, stderr=PIPE) output = p.communicate()[0].splitlines() if len(output) > 2: sizes = output[-1].split()[0:4] self.text = int(sizes[0]) self.data = int(sizes[1]) self.bss = int(sizes[2]) self.total = int(sizes[3]) else: self.text = self.data = self.bss = self.total = 0 def show(self, indent=""): print "%-32s %10d | %10d %10d %10d" % \ (indent+self.title, self.total, self.text, self.data, self.bss) class Report: def create(filename, title, subglob=None): r = Report(filename, title) path = os.path.dirname(filename) p = Popen("ls " + path + "/*.o | grep -v built-in.o", shell=True, stdout=PIPE, stderr=PIPE) glob = join(p.communicate()[0].splitlines()) oreport = Report(glob, path + "/*.o") oreport.sizes.title = path + "/*.o" r.parts.append(oreport) if subglob: p = Popen("ls " + subglob, shell=True, stdout=PIPE, stderr=PIPE) for f in p.communicate()[0].splitlines(): path = os.path.dirname(f) r.parts.append(Report.create(f, path, path + "/*/built-in.o")) r.parts.sort(reverse=True) for b in r.parts: r.totals["total"] += b.sizes.total r.totals["text"] += b.sizes.text r.totals["data"] += b.sizes.data r.totals["bss"] += b.sizes.bss r.deltas["total"] = r.sizes.total - r.totals["total"] r.deltas["text"] = r.sizes.text - r.totals["text"] r.deltas["data"] = r.sizes.data - r.totals["data"] r.deltas["bss"] = r.sizes.bss - r.totals["bss"] return r create = staticmethod(create) def __init__(self, glob, title): self.glob = glob self.title = title self.sizes = Sizes(glob) self.parts = [] self.totals = {"total":0, "text":0, "data":0, "bss":0} self.deltas = {"total":0, "text":0, "data":0, "bss":0} def show(self, indent=""): rule = str.ljust(indent, 80, '-') print "%-32s %10s | %10s %10s %10s" % \ (indent+self.title, "total", "text", "data", "bss") print rule self.sizes.show(indent) print rule for p in self.parts: if p.sizes.total > 0: p.sizes.show(indent) print rule print "%-32s %10d | %10d %10d %10d" % \ (indent+"sum", self.totals["total"], self.totals["text"], self.totals["data"], self.totals["bss"]) print "%-32s %10d | %10d %10d %10d" % \ (indent+"delta", self.deltas["total"], self.deltas["text"], self.deltas["data"], self.deltas["bss"]) print "\n" def __cmp__(this, that): if that is None: return 1 if not isinstance(that, Report): raise TypeError if this.sizes.total < that.sizes.total: return -1 if this.sizes.total > that.sizes.total: return 1 return 0 def main(): try: opts, args = getopt.getopt(sys.argv[1:], "dh", ["help"]) except getopt.GetoptError, err: print '%s' % str(err) usage() sys.exit(2) driver_detail = False for o, a in opts: if o == '-d': driver_detail = True elif o in ('-h', '--help'): usage() sys.exit(0) else: assert False, "unhandled option" glob = "arch/*/built-in.o */built-in.o" vmlinux = Report.create("vmlinux", "Linux Kernel", glob) vmlinux.show() for b in vmlinux.parts: if b.totals["total"] > 0 and len(b.parts) > 1: b.show() if b.title == "drivers" and driver_detail: for d in b.parts: if d.totals["total"] > 0 and len(d.parts) > 1: d.show(" ") if __name__ == "__main__": main() ########NEW FILE########
UTF-8
Python
false
false
1,724,689
py
16,703
allPythonContent.py
3,862
0.568044
0.560071
0.000032
46,342
35.216542
335
Weevils/India_Climate
4,990,752,020,632
97a7b3ffda3d790323ab200f4040e93a1d933e6e
0c22e50b81bc59bc3bed5eb0d3339964d50f2b47
/Scripts/DataAnalysis_Yield_Stability_Quick.py
4361529e4747281dc4a199f9c68e0e4c7cf9cd0b
[]
no_license
https://github.com/Weevils/India_Climate
698941d36b38a244887e1453dd9901d800462ac3
a5fb70b88894211f2db36bcf4cbb700aee9e0821
refs/heads/master
2019-08-20T20:48:07.954099
2016-09-23T22:57:43
2016-09-23T22:57:52
58,396,671
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
__author__ = 'Vinny_Ricciardi' #Goal: merge and work with yield, climate (tmp, pre), and farm size data # Todo: expand to tmp(erature), for now just pre(cipitation) import pandas as pd from datetime import datetime import os import glob import matplotlib as plt import seaborn as sns import numpy as np import sys import pandas as pd import numpy as np pd.options.display.float_format = '{:.2f}'.format import seaborn as sns; sns.set(color_codes=True) sns.set() import glob import os import time import matplotlib.pyplot as plt plt.style.use('ggplot') from mappy.utilities.plot_map import plot_map startTime = datetime.now() print startTime # Step 1: Prep data # Read in and merge data: precipitation, farm size, yields TEMP = '/Users/Vinny_Ricciardi/Downloads/test.csv' PATH_VDSA = '/Users/Vinny_Ricciardi/Documents/Data_Library/Survey/India/India_VDSA/' \ 'Country_Wide_merged_timescales/temporal/' PATH_CRU = '/Users/Vinny_Ricciardi/Documents/Data_Library/Survey/India/India_VDSA/' \ 'CRU_Climate_VDSA_DIST/CRU_Cleaned_VDSA/' base_df = pd.read_csv(PATH_VDSA + 'dt_operational_holdings_a_web.csv') #clean farm size data base_df = base_df.replace('-1', np.nan, regex=True) base_df['YEAR'] = base_df['YEAR'].replace(1974, 1975) base_df['YEAR'] = base_df['YEAR'].replace(1976, 1975) base_df['YEAR'] = base_df['YEAR'].replace(1979, 1980) base_df['YEAR'] = base_df['YEAR'].replace(1986, 1985) base_df['YEAR'] = base_df['YEAR'].replace(2001, 2000) #merge base with in a df not to be used, but will enable interpolation b/c contains all years needed dummy = pd.read_csv(PATH_VDSA + 'dt_market_road_a_web.csv') cv = ['DIST', 'YEAR'] base_df = pd.merge(base_df, dummy, how='outer', on=cv) base_df = base_df.sort_values(['YEAR']) district_list = [base_df.query('DIST == {}'.format(dist)).interpolate(method='linear') for dist in base_df.DIST.unique()] base_df = pd.concat(district_list) # merge climate and yield data to newly interpolated farm size data def merge_VDSA(base_df, df_list): for value in df_list: base_df = pd.merge(base_df, value, how='outer', on=cv) return base_df df_list = [ pd.read_csv(PATH_VDSA + 'dt_area_prod_a_web.csv'), pd.read_csv(PATH_CRU + 'Cleaned_.pre_all_MEAN.csv') ] df = merge_VDSA(base_df, df_list) df = df.replace('-1', np.nan, regex=True) df.columns # Calculate percent farm size per category; note, TOTAL_NO is new because VDSA version didn't aways equal 100 percent df['TOTAL_NO'] = df[['MARG_NO', 'SMALL_NO', 'SEMIM_NO', 'MEDIUM_NO', 'LARGE_NO']].sum(axis=1) df['MARG_NP'] = df['MARG_NO']/df['TOTAL_NO'] df['SMALL_NP'] = df['SMALL_NO']/df['TOTAL_NO'] df['SEMIM_NP'] = df['SEMIM_NO']/df['TOTAL_NO'] df['MEDIUM_NP'] = df['MEDIUM_NO']/df['TOTAL_NO'] df['LARGE_NP'] = df['LARGE_NO']/df['TOTAL_NO'] df['TOTAL_NP_Test'] = df[['MARG_NP', 'SMALL_NP', 'SEMIM_NP', 'MEDIUM_NP', 'LARGE_NP']].sum(axis=1) df['TOTAL_NP_Test'].describe() df = df.query('TOTAL_NP_Test > 0') # check out Farm_Size_EDA.py for percentage plots if needed # add to this .py for fixed interpolation # save df to csv for panel analysis df.rename(columns={'SCAN_TA': 'SGUR_TA'}, inplace=True) df.to_csv('/Users/Vinny_Ricciardi/Documents/Data_Library/Survey/India/India_VDSA/VDSA_lm_ready/df_panel.csv') ############################################################################################################ # Step 2. Prepare df for analysis: cv pre and yield, median farm size percentages # clean production variables # clean data: presently missing from below list # df.rename(columns={'SCAN_TA': 'SGUR_TA'}, inplace=True) #sugar # df.rename(columns={'SORG_KA': 'SORG_K_TA'}, inplace=True) #karif sorghum production # df.rename(columns={'SORG_KQ': 'SORG_K_TQ'}, inplace=True) #karif sorghum area # df.rename(columns={'SORG_RA': 'SORG_R_TA'}, inplace=True) #rabi sorghum production # df.rename(columns={'SORG_RQ': 'SORG_R_TQ'}, inplace=True) #rabi sorghum area grouped = df.groupby('DIST') # Calculate CV precipitation (sigma over mu) for annual, karif and rabi # Todo: make generalizable for all months # Todo: make generalizable for al climate var (esp tmp) # annual gr_annual = grouped['annual'].agg({'annual_avg': np.mean, 'annual_std': np.std}) gr_annual['annual_cov'] = gr_annual['annual_std']/gr_annual['annual_avg'] gr_annual.reset_index(inplace=True) # rabi gr_rab = grouped['rabi'].agg({'rabi_avg': np.mean, 'rabi_std': np.std}) gr_rab['rabi_cov'] = gr_rab['rabi_std']/gr_rab['rabi_avg'] gr_rab.reset_index(inplace=True) # karif gr_karif = grouped['karif'].agg({'karif_avg': np.mean, 'karif_std': np.std}) gr_karif['karif_cov'] = gr_karif['karif_std']/gr_karif['karif_avg'] gr_karif.reset_index(inplace=True) #merge pre gr_pre = pd.merge(gr_annual, gr_rab, how='outer', on='DIST') gr_pre = pd.merge(gr_karif, gr_pre, how='outer', on='DIST') # Calculate CV yield for rice, cotton, sugar #Todo: generalize to all crops df['RICE_YIELD'] = df['RICE_TQ']/df['RICE_TA'] gr_ri = grouped['RICE_YIELD'].agg({'RICE_avg': np.mean, 'RICE_std': np.std}) gr_ri['RICE_cov'] = gr_ri['RICE_std']/gr_ri['RICE_avg'] gr_ri.reset_index(inplace=True) df['COTN_YIELD'] = df['COTN_TQ']/df['COTN_TA'] gr_ct = grouped['COTN_YIELD'].agg({'COTN_avg': np.mean, 'COTN_std': np.std}) gr_ct['COTN_cov'] = gr_ct['COTN_std']/gr_ct['COTN_avg'] gr_ct.reset_index(inplace=True) df['SGUR_YIELD'] = df['SGUR_TQ']/df['SGUR_TA'] gr_sg = grouped['SGUR_YIELD'].agg({'SGUR_avg': np.mean, 'SGUR_std': np.std}) gr_sg['SGUR_cov'] = gr_sg['SGUR_std']/gr_sg['SGUR_avg'] gr_sg.reset_index(inplace=True) gr_yd = pd.merge(gr_ri, gr_ct, how='outer', on='DIST') gr_yd = pd.merge(gr_sg, gr_yd, how='outer', on='DIST') # Calculate median farm size proportions over time # Todo: check other stats # Todo: check percent area too gr_fm = df.loc[:,['DIST', 'LARGE_NP', 'MEDIUM_NP', 'SEMIM_NP', 'SMALL_NP', 'MARG_NP']] gr_fm = gr_fm.groupby('DIST').median() gr_fm.reset_index(inplace=True) # Merge 3 groups back to common df df = pd.merge(gr_pre, gr_yd, how='outer', on='DIST') df = pd.merge(gr_fm, df, how='outer', on='DIST') df.columns df.head() df.to_csv('/Users/Vinny_Ricciardi/Documents/Data_Library/Survey/India/India_VDSA/VDSA_lm_ready/AAG_lm_input.csv') #Todo: keep for crop generilization: # clean data: presently missing from below list # df.rename(columns={'SCAN_TA': 'SGUR_TA'}, inplace=True) #sugar # df.rename(columns={'SORG_KA': 'SORG_K_TA'}, inplace=True) #karif sorghum production # df.rename(columns={'SORG_KQ': 'SORG_K_TQ'}, inplace=True) #karif sorghum area # df.rename(columns={'SORG_RA': 'SORG_R_TA'}, inplace=True) #rabi sorghum production # df.rename(columns={'SORG_RQ': 'SORG_R_TQ'}, inplace=True) #rabi sorghum area # crops = ['BRLY', 'CAST', 'CERL', 'COTN', 'CPEA', 'FMLT', 'GNUT', 'LINS', # 'MAIZ', 'OILS', 'PMLT', 'PPEA', 'PULS', 'RICE', 'RM', 'SAFF', # 'SESA', 'SORG', 'SOYA', 'SUNF', 'WHT'] ############################################################################################################ #Todo below is very messy! # Plotting # Correlation plot # Compute the correlation matrix # corr = df.corr() # # # Generate a mask for the upper triangle # mask = np.zeros_like(corr, dtype=np.bool) # mask[np.triu_indices_from(mask)] = True # # # Set up the matplotlib figure # f, ax = plt.subplots(figsize=(11, 9)) # # # Generate a custom diverging colormap # cmap = sns.diverging_palette(220, 10, as_cmap=True) # # # yticks = df.index # # keptticks = yticks[::int(len(yticks)/10)] # # yticks = ['' for y in yticks] # # yticks[::int(len(yticks)/10)] = keptticks # # # # xticks = df.columns # # keptticks = xticks[::int(len(xticks)/10)] # # xticks = ['' for y in xticks] # # xticks[::int(len(xticks)/10)] = keptticks # # # # Draw the heatmap with the mask and correct aspect ratio # sns.heatmap(corr, mask=mask, cmap=cmap, vmax=.3, annot=True, # square=True, xticklabels=5, yticklabels=5, # linewidths=.5, cbar_kws={"shrink": .5}, ax=ax) # # plt.yticks(rotation=0) # # plt.show() # # # # # Step 3: bivariate plots farm size yield # fig1 = plt.figure(1, figsize=(25,10)) # # fig1.suptitle('%s' % x, fontsize=20) # # z = (4,5) # plot size dimensions # # # row 1: number of farmer distribution per size class # ax1 = plt.subplot2grid(z, (0,0), rowspan=1, colspan=1) # ax2 = plt.subplot2grid(z, (0,1), rowspan=1, colspan=1) # ax3 = plt.subplot2grid(z, (0,2), rowspan=1, colspan=1) # ax4 = plt.subplot2grid(z, (0,3), rowspan=1, colspan=1) # ax5 = plt.subplot2grid(z, (0,4), rowspan=1, colspan=1) # # ax1.yaxis.grid(False) # ax2.yaxis.grid(False) # ax3.yaxis.grid(False) # ax4.yaxis.grid(False) # ax5.yaxis.grid(False) # # ax1.xaxis.grid(False) # ax2.xaxis.grid(False) # ax3.xaxis.grid(False) # ax4.xaxis.grid(False) # ax5.xaxis.grid(False) # # ax1.xaxis.set_ticklabels([]) # ax2.xaxis.set_ticklabels([]) # ax3.xaxis.set_ticklabels([]) # ax4.xaxis.set_ticklabels([]) # ax5.xaxis.set_ticklabels([]) # # ax1.set_title('Marginal', fontsize=12) # ax2.set_title('Small', fontsize=12) # ax3.set_title('SemiMedium', fontsize=12) # ax4.set_title('Medium', fontsize=12) # ax5.set_title('Large', fontsize=12) # # sns.kdeplot(df.MARG_NP, ax=ax1, label='') # sns.kdeplot(df.SMALL_NP, ax=ax2, label='') # sns.kdeplot(df.SEMIM_NP, ax=ax3, label='') # sns.kdeplot(df.MEDIUM_NP, ax=ax4, label='') # sns.kdeplot(df.LARGE_NP, ax=ax5, label='') # # # # Row 2: Bivariate farm size yield relationships # ax12 = plt.subplot2grid(z, (1,0), rowspan=1, colspan=1) # ax22 = plt.subplot2grid(z, (1,1), rowspan=1, colspan=1) # ax32 = plt.subplot2grid(z, (1,2), rowspan=1, colspan=1) # ax42 = plt.subplot2grid(z, (1,3), rowspan=1, colspan=1) # ax52 = plt.subplot2grid(z, (1,4), rowspan=1, colspan=1) # # ax12.yaxis.grid(False) # # ax12.xaxis.set_ticklabels([]) # ax22.xaxis.set_ticklabels([]) # ax32.xaxis.set_ticklabels([]) # ax42.xaxis.set_ticklabels([]) # ax52.xaxis.set_ticklabels([]) # # ax12.axes.get_xaxis().set_visible(False) # ax22.axes.get_xaxis().set_visible(False) # ax32.axes.get_xaxis().set_visible(False) # ax42.axes.get_xaxis().set_visible(False) # ax52.axes.get_xaxis().set_visible(False) # # ax22.axes.get_yaxis().set_visible(False) # ax32.axes.get_yaxis().set_visible(False) # ax42.axes.get_yaxis().set_visible(False) # ax52.axes.get_yaxis().set_visible(False) # # sns.regplot(df['MARG_NP'], df['RICE_cov'], ax=ax12) # sns.regplot(df['SMALL_NP'], df['RICE_cov'], ax=ax22) # sns.regplot(df['SEMIM_NP'], df['RICE_cov'], ax=ax32) # sns.regplot(df['MEDIUM_NP'], df['RICE_cov'], ax=ax42) # sns.regplot(df['LARGE_NP'], df['RICE_cov'], ax=ax52) # # # Row 3: Bivariate farm size to climate relationships # ax12 = plt.subplot2grid(z, (2,0), rowspan=1, colspan=1) # ax22 = plt.subplot2grid(z, (2,1), rowspan=1, colspan=1) # ax32 = plt.subplot2grid(z, (2,2), rowspan=1, colspan=1) # ax42 = plt.subplot2grid(z, (2,3), rowspan=1, colspan=1) # ax52 = plt.subplot2grid(z, (2,4), rowspan=1, colspan=1) # # ax12.yaxis.grid(False) # # ax12.xaxis.set_ticklabels([]) # ax22.xaxis.set_ticklabels([]) # ax32.xaxis.set_ticklabels([]) # ax42.xaxis.set_ticklabels([]) # ax52.xaxis.set_ticklabels([]) # # ax12.axes.get_xaxis().set_visible(False) # ax22.axes.get_xaxis().set_visible(False) # ax32.axes.get_xaxis().set_visible(False) # ax42.axes.get_xaxis().set_visible(False) # ax52.axes.get_xaxis().set_visible(False) # # ax22.axes.get_yaxis().set_visible(False) # ax32.axes.get_yaxis().set_visible(False) # ax42.axes.get_yaxis().set_visible(False) # ax52.axes.get_yaxis().set_visible(False) # # sns.regplot(df['MARG_NP'], df['annual_cov'], ax=ax12) # sns.regplot(df['SMALL_NP'], df['annual_cov'], ax=ax22) # sns.regplot(df['SEMIM_NP'], df['annual_cov'], ax=ax32) # sns.regplot(df['MEDIUM_NP'], df['annual_cov'], ax=ax42) # sns.regplot(df['LARGE_NP'], df['annual_cov'], ax=ax52) # # # # Row 4: Bivariate climate to yield relationships # ax14 = plt.subplot2grid(z, (3,0), rowspan=1, colspan=1) # ax24 = plt.subplot2grid(z, (3,1), rowspan=1, colspan=1) # ax34 = plt.subplot2grid(z, (3,2), rowspan=1, colspan=1) # # ax14.yaxis.grid(False) # # ax14.xaxis.set_ticklabels([]) # ax24.xaxis.set_ticklabels([]) # ax34.xaxis.set_ticklabels([]) # # # ax14.axes.get_xaxis().set_visible(False) # # ax24.axes.get_xaxis().set_visible(False) # # ax34.axes.get_xaxis().set_visible(False) # # ax24.axes.get_yaxis().set_visible(False) # ax34.axes.get_yaxis().set_visible(False) # # sns.regplot(df['annual_cov'], df['RICE_cov'], ax=ax14) # sns.regplot(df['rabi_cov'], df['RICE_cov'], ax=ax24) # sns.regplot(df['karif_cov'], df['RICE_cov'], ax=ax34) # print datetime.now() - startTime # # # # #relationships # #farm size and yield # #yield and climate # #farm size and climate # # fig1 = plt.figure(1, figsize=(20,3)) # # fig1.suptitle('%s' % x, fontsize=20) # # z = (1,5) # plot size dimensions # # ax1 = plt.subplot2grid(z, (0,0), rowspan=1, colspan=1) # ax2 = plt.subplot2grid(z, (0,1), rowspan=1, colspan=1) # ax3 = plt.subplot2grid(z, (0,2), rowspan=1, colspan=1) # ax4 = plt.subplot2grid(z, (0,3), rowspan=1, colspan=1) # ax5 = plt.subplot2grid(z, (0,4), rowspan=1, colspan=1) # # sns.regplot(df['MARG_NP'], df['RICE_cov'], ax=ax1) # sns.regplot(df['SMALL_NP'], df['RICE_cov'], ax=ax2) # sns.regplot(df['SEMIM_NP'], df['RICE_cov'], ax=ax3) # sns.regplot(df['MEDIUM_NP'], df['RICE_cov'], ax=ax4) # sns.regplot(df['LARGE_NP'], df['RICE_cov'], ax=ax5) # # #mapping # # shp = '/Users/Vinny_Ricciardi/Documents/Data_Library/GIS_Library/India/India_Admin_shp/Dist_1970_VDSA/District_1970_Dict_proj.shp' # PATH_1 = '/Users/Vinny_Ricciardi/Downloads/' # # # m = df['RICE_cov'].max() # # # for i in x: # plot_map(shp, df, unique_shp=['DIST'], unique_tbl=['DIST'], variable2map= ['RICE_cov'], # min_value=0., max_value=m, fig_size=(2, 3), dpi=200, projection='lcc', # out_figure= '/Users/Vinny_Ricciardi/Downloads/Map_cov_test.png') # # help(plot_map) # # print datetime.now() - startTime
UTF-8
Python
false
false
14,086
py
10
DataAnalysis_Yield_Stability_Quick.py
9
0.647096
0.617919
0
426
32.068075
132
yeldiRium/2048
11,587,821,788,200
b79bf649a4fff85a2cbc7361546a9248f7d56b80
c6895e2b1e81ad5ad7309c379c35a1e533b8570d
/exceptions.py
1a2fc21b335099eca214b34374801889fcbfa1e2
[]
no_license
https://github.com/yeldiRium/2048
f9d8f85eab141f430df2116bc053633902b19713
8e12b3a4140d93c51a783914bfabdeb2682e7679
refs/heads/master
2021-01-21T12:59:08.329248
2016-04-09T23:06:44
2016-04-09T23:06:44
55,513,557
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
class GameNotInitializedError(Exception): pass class GameLostError(Exception): pass class InvalidActionError(Exception): pass class NoEmptyContainerError(Exception): pass class MoveNotAllowedError(Exception): pass class FusionNotAllowedError(Exception): pass
UTF-8
Python
false
false
293
py
23
exceptions.py
22
0.761092
0.761092
0
22
12.318182
41
Mfhodges/CRF2
19,679,540,155,105
4e1823a344d604e9c1a5b20b22aa07722afaedcd
e3095c9f598ad7c59a059ca27554546c18906863
/crf2/custom_auth.py
1109e6c75b7286a5480d8c8f96c8eb841dea3f2c
[]
no_license
https://github.com/Mfhodges/CRF2
9d6c465fabad7c7e09b8324675a460b12e9f71fa
636f17ed540456ec867c17f8078dd96ffcdbabc4
refs/heads/master
2022-02-27T23:06:45.884356
2020-10-12T18:31:43
2020-10-12T18:31:43
172,382,692
1
3
null
null
null
null
null
null
null
null
null
null
null
null
null
from django.contrib.auth.middleware import RemoteUserMiddleware from django.contrib.auth.models import User from course.models import Profile class CustomRemoteUserMiddleware(RemoteUserMiddleware): def configure_user(request, user): meta = request.META try: user.update(first_name=meta['givenName'],last_name=meta['sn'],email=meta['mail']) Profile.objects.update_or_create(user=user,penn_id=meta['penn_id']) except: #fail silently pass
UTF-8
Python
false
false
514
py
123
custom_auth.py
56
0.684825
0.684825
0
14
35.714286
93
jack-mars/selemium_webdriver
1,503,238,596,648
4b222385bcb82302332228ca379442a7a6b7b74b
e9e11aee109878d490a77eef74fb14aa1538f684
/test37understandingmethods.py
071bc8082bc8aa719d0b1d016ed76219766b76ec
[]
no_license
https://github.com/jack-mars/selemium_webdriver
c107990a1564ec038db6b7c52556fd7438aa13f2
1e76ac27a47422bc996831946974a3c107202ae8
refs/heads/master
2020-04-01T09:00:03.503760
2018-10-25T03:55:18
2018-10-25T03:55:18
153,055,658
0
0
null
false
2018-10-16T05:54:12
2018-10-15T05:03:15
2018-10-15T17:28:33
2018-10-16T05:54:11
20
0
0
0
Python
false
null
def sum_nums(n1, n2): print(n1 + n2) sum_nums(2, 8) l = [1, 2, 3, 2, 2] print(l.count(2)) print(l.append(6)) print(l) print(len(l))
UTF-8
Python
false
false
135
py
50
test37understandingmethods.py
47
0.577778
0.481481
0
8
16
21
Dachoka3000/password-locker
2,611,340,151,366
01ac8183a595b9e117d0b3b630204afcd67152e1
3221080135f2740a29738144b52e18d026347c5f
/credentials_test.py
971e355cfcd96c1a7f2177bf03ef5d8a93d12740
[ "MIT" ]
permissive
https://github.com/Dachoka3000/password-locker
82aeffcebf9198a492d2107cffaff70b1d47474a
179f0c29c03b0a20f09a477f1b4ef86ab7d18c27
refs/heads/master
2022-12-22T15:47:21.084424
2020-09-16T15:29:00
2020-09-16T15:29:00
null
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
import unittest import pyperclip from credentials import Credentials class TestCredentials(unittest.TestCase): ''' Test class that defines test cases for the credentials class behaviours Args: unittest.TestCase: TestCase class that helps in creating test cases ''' def tearDown(self): ''' tearDown method that cleans up after each tesst case is run ''' Credentials.credentials_list = [] def setUp(self): ''' Setup method to run before each test cases ''' self.new_credentials = Credentials("Twitter", "MaryX", "maryx1000") def test_init(self): ''' test_init test case to test if the object is initialised correctly ''' self.assertEqual(self.new_credentials.account, "Twitter") self.assertEqual(self.new_credentials.user_name, "MaryX") self.assertEqual(self.new_credentials.pass_word,"maryx1000") def test_save_credentials(self): ''' test_save_credentials to test if the credentials objects are saved into the credentials list ''' self.new_credentials.save_credentials() self.assertEqual(len(Credentials.credentials_list), 1) def test_save_multiple_credentials(self): ''' test_save_multiple_credentials to check if we can saave multiple credentials ''' self.new_credentials.save_credentials() test_credential = Credentials("Instagram", "Muse Art", "museart1000") test_credential.save_credentials() self.assertEqual(len(Credentials.credentials_list), 2) def test_delete_credentials(self): ''' test to check if we can delete credentials from the credentials list ''' self.new_credentials.save_credentials() test_credential = Credentials("Instagram", "Muse Art", "museart1000") test_credential.save_credentials() self.new_credentials.delete_credentials() self.assertEqual(len(Credentials.credentials_list),1) def test_find_credentials_by_account(self): ''' test to check if we can find credentials by account and display information ''' self.new_credentials.save_credentials() test_credential = Credentials("Snapchat", "Test User", "user195") test_credential.save_credentials() found_credential = Credentials.find_by_account("Snapchat") self.assertEqual(found_credential.pass_word, test_credential.pass_word) def test_credentials_exists(self): ''' test to check if we can return a boolean if we cannot find the credentials ''' self.new_credentials.save_credentials() test_credential = Credentials("Snapchat", "Test User", "user195") test_credential.save_credentials() credential_exists = Credentials.credentials_exist("Snapchat") self.assertTrue(credential_exists) def test_display_all_credentials(self): ''' method that returns a list of all credentials saved ''' self.assertEqual(Credentials.display_credentials(), Credentials.credentials_list) def test_copy_details(self): ''' test to confirm that we are copying the details from a found account ''' self.new_credentials.save_credentials() Credentials.copy_details("Snapchat") self.assertEqual(self.new_credentials.user_name, pyperclip.paste()) if __name__ == '__main__': unittest.main()
UTF-8
Python
false
false
3,515
py
6
credentials_test.py
5
0.652916
0.645804
0
109
31.229358
100
shikouchen/WalkingTKK_crawler
7,198,365,199,514
cac43cce1e43cff4b8d3f2127b0ad9e321729e3a
43c0be39103305ad8ad6b8fee78b3d515a13f8de
/jw_update_info/check_ele_info.py
f39ee281f35db83d8b2ef30a8c8fb84132d8bc6f
[]
no_license
https://github.com/shikouchen/WalkingTKK_crawler
96b66b22a09801e769684fcdf51c365efdac1e64
6e3567c74901575d446dad0c4fe50368fb567796
refs/heads/master
2020-05-31T16:59:12.824500
2019-06-05T13:09:23
2019-06-05T13:09:23
null
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
#!/usr/bin/env python3 # -*- coding:utf-8 -*- # Author: Erick from jw_info_update_method import * from log import logger import os,sys import urllib import xml.dom.minidom import time import datetime import json import urllib.request import requests class CheckEleInfo(): def __init__(self): pass def ReadSet(self): host = "" user = "" passwd = "" db = "" charset = "" if os.path.exists("settings.conf"): dom = xml.dom.minidom.parse('settings.conf') data = dom.documentElement try: host = data.getElementsByTagName("host")[0].childNodes[0].data user = data.getElementsByTagName("user")[0].childNodes[0].data passwd = data.getElementsByTagName("passwd")[0].childNodes[0].data db = data.getElementsByTagName("db")[0].childNodes[0].data charset = data.getElementsByTagName("charset")[0].childNodes[0].data except: print("The configuration information is incomplete") sys.exit(0) return host,user,passwd,db,charset def check_ele(self): host = "120.78.63.47" user = "api" passwd = "forsource,110" db = "tkk_app" charset = 'utf8' print("更新电费信息--开始--" + str(datetime.datetime.now())) # host, user, passwd, db, charset = self.ReadSet() print(host) jw_update = JW(host, user, passwd, db, charset) url = 'https://www.xujc.club/api/userallinfofromtkkdatabase' headers = { "Authorization": "Basic YWRtaW46YWRtaW4xMjM0NQ==" } request = urllib.request.Request(url, headers=headers) rawData = urllib.request.urlopen(request) jsonStr = rawData.read() # 使用read方法将json数据读出 users = json.loads(jsonStr.decode('utf8')) # 将json数据转换为python的列表格式 # print(users) for user in users: tel = user["tel"] drRoomId = user["drRoomId"] drlou = user["drlou"] drxiaoqu = user["drxiaoqu"] send_message_status = user["send_message_status"] if drxiaoqu and drlou and drRoomId: try: ele_url = "https://www.xujc.club/api/electricchargeinfo/" + tel ele_resp = requests.get(url=ele_url, headers=headers) ele_raw_data = ele_resp.text ele_raw_data = json.loads(ele_raw_data) # 将json数据转换为python的列表格式 result = ele_raw_data[0]["result"] if result != "1": print(tel + "---userinfo表中的宿舍信息:" + drxiaoqu + drlou + drRoomId) try: jw_update.get_ele(tel, drxiaoqu, drlou, drRoomId, send_message_status) except Exception as e: print("crawl info--更新电费信息失败--" + str(datetime.datetime.now()) + "--" + tel + "---" + str(e)) except Exception as e: print("获取电费信息失败" + tel + "---" + str(e)) print("更新电费信息--结束") if __name__ == '__main__': check = CheckEleInfo() check.check_ele()
UTF-8
Python
false
false
3,351
py
30
check_ele_info.py
22
0.539732
0.529448
0
85
36.705882
120
lukewbecker/classification-exercises
9,835,475,125,690
0ed5c1582374c7513bcd4e8716e2848ded2072e8
9f0a648ace0555f5a7af3a7a59a4b34c74f8238e
/aquire.py
651f4048638f9d333ac30e24e88a3826e43531eb
[]
no_license
https://github.com/lukewbecker/classification-exercises
10afb485a517d814df108abc718d8505f9c19d42
fa21ad6e2466896d2a945ed8b2fe522b35487183
refs/heads/master
2022-12-12T05:59:21.089591
2020-09-24T13:24:26
2020-09-24T13:24:26
293,915,453
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
# Creating the aquire.py file # Make a new python module, acquire.py to hold the following data aquisition functions: # get_titanic_data # get_iris_data # Importing libraries: import pandas as pd import matplotlib.pyplot as plt import numpy as np import seaborn as sns # Importing the os library specifically for reading the csv once I've created the file in my working directory. import os # Make a function named get_titanic_data that returns the titanic data from the codeup data science database as a pandas data frame. Obtain your data from the Codeup Data Science Database. # Setting up the user credentials: from env import host, user, password def get_db(db, user=user, host=host, password=password): return f'mysql+pymysql://{user}:{password}@{host}/{db}' # titanic_db query sql_query = 'SELECT * FROM passengers' def get_titanic_data(): return pd.read_sql('SELECT * FROM passengers', get_db('titanic_db')) # Showing the df: # print(get_titanic_data()) # Make a function named get_iris_data that returns the data from the iris_db on the codeup data science database as a pandas data frame. The returned data frame should include the actual name of the species in addition to the species_ids. Obtain your data from the Codeup Data Science Database. def get_iris_data(): return pd.read_sql('''SELECT measurement_id, sepal_length, sepal_width, petal_length, petal_width, m.species_id, species_name FROM measurements AS m JOIN species AS s on m.species_id = s.species_id;''', get_db('iris_db')) # Once you've got your get_titanic_data and get_iris_data functions written, now it's time to add caching to them. To do this, edit the beginning of the function to check for a local filename like titanic.csv or iris.csv. If they exist, use the .csv file. If the file doesn't exist, then produce the SQL and pandas necessary to create a dataframe, then write the dataframe to a .csv file with the appropriate name. def get_titanic_data_cache(): filename = 'titanic.csv' if os.path.isfile(filename): return pd.read_csv(filename) else: titanic_df = pd.read_sql('SELECT * FROM passengers', get_db('titanic_db')) titanic_df.to_csv(filename, index = False) # Ryan was using .to_file, but I was getting error when trying to use that function. return titanic_df # print(get_titanic_data_cache()) # Defining the variable for my titanic data: # df_titanic = get_titanic_data_cache() # # Creating a seaborn chart: # sns.relplot(data = df_titanic, x = 'age', y = 'fare', hue = 'class') # plt.show() # Doing the same for the iris_db: def get_iris_data_cache(): filename = 'iris.csv' if os.path.isfile(filename): return pd.read_csv(filename) else: iris_df = get_iris_data() iris_df.to_csv(filename, index = False) # Ryan was using .to_file, but I was getting error when trying to use that function. return iris_df # print(type(get_iris_data_cache())) # print(get_iris_data_cache()) print('End of file.')
UTF-8
Python
false
false
3,041
py
10
aquire.py
4
0.713252
0.713252
0
82
36.097561
414
HelenProkopova/structural-programming
14,010,183,349,418
8cf36a42a6bb33e9aea28671176b2c3c0fe32fc8
560549f12e4139d31225d806406d489ec9b36f0c
/4.1.py
3dc2f2229adaf0a7a42e1d55fd05ef6816b79f81
[]
no_license
https://github.com/HelenProkopova/structural-programming
ee937f5072efb80594d53a5c1131e0de88800be5
56f9885dbf34024a4da2d038110a50c002d8f279
refs/heads/master
2021-01-19T14:40:45.800619
2017-05-24T18:43:48
2017-05-24T18:43:48
88,181,810
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
s=input ('введите строку: ') t=int(0) for i in range(len(s)): if s[i]== ',' or s[i]== '.' : t=t+1 print ('количество точек и запятых= ',t)
UTF-8
Python
false
false
212
py
25
4.1.py
24
0.448864
0.4375
0
7
20.714286
40
katarzynalatos/Lab_1
7,533,372,642,781
5b27a787979b1b8249feed8d02b9e52c06aff541
9d1a857b529e2c1fe9ae4174b5d859e094299a86
/problem2/InputFileValidator.py
1c9be451969cac835cc9f3af4d210facb03f399e
[]
no_license
https://github.com/katarzynalatos/Lab_1
900c40e2513be858cceda8a7b9b20325fe30b029
710bcd135775783337c5728c88d8283e800419b9
refs/heads/master
2020-05-25T23:59:03.156302
2017-04-05T06:53:38
2017-04-05T06:53:38
84,980,182
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
from os.path import isfile from os.path import isdir from os import listdir from AbstractValidator import AbstractValidator class InputFileValidator(AbstractValidator): def __init__(self, name): self._name=name self._files = [] def validate(self): if isfile(self._name): self._files.append(self._name) return True elif isdir(self._name): self._files = [self._name+"/"+files for files in listdir(self._name) if isfile(self._name+"/"+files)] if len(self._files) > 0: return True else: return False else: return False def get_list_of_files(self): return self._files
UTF-8
Python
false
false
757
py
7
InputFileValidator.py
5
0.557464
0.556143
0
26
27.115385
113
Aasthaengg/IBMdataset
3,848,290,739,084
d6999daea8ff69ecd418869fce153499d0e9368d
ca7aa979e7059467e158830b76673f5b77a0f5a3
/Python_codes/p02747/s241428742.py
c8aff520471c477e8a417c13c4623a03de673779
[]
no_license
https://github.com/Aasthaengg/IBMdataset
7abb6cbcc4fb03ef5ca68ac64ba460c4a64f8901
f33f1c5c3b16d0ea8d1f5a7d479ad288bb3f48d8
refs/heads/main
2023-04-22T10:22:44.763102
2021-05-13T17:27:22
2021-05-13T17:27:22
367,112,348
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
s=input() if len(s)%2==0: for i in range(len(s)//2): if s[2*i:2*i+2]=="hi": continue print("No") break else: print("Yes") else: print("No")
UTF-8
Python
false
false
167
py
202,060
s241428742.py
202,055
0.497006
0.461078
0
11
14.272727
28
ericw13/radabo
2,138,893,759,898
925b67370ff7768675d691992373fdb5a4061780
463aaa1a47c50f93e19e7b6b7d4d00bbdaec7769
/app/radabo/urls.py
7a9971749ef4ecc0ef9890ec71139dada65db41b
[]
no_license
https://github.com/ericw13/radabo
83bee8d1c1dadac05999876a69d8a650f97dcfce
ed6b9ba4b57d8fccc59686ea66c98ed23e9646f0
refs/heads/master
2021-01-12T06:55:25.197106
2016-12-19T15:09:50
2016-12-19T15:09:50
76,681,084
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
from django.conf.urls import url from radabo import views urlpatterns = [ url(r'^$', views.IndexView.as_view(), name='index'), url(r'^prioritization/?$', views.Priority, name='priority'), url(r'^dashboard/?$', views.Dashboard, name='dashboard'), url(r'^enhancements/release/?$', views.ReleaseReport, name='release'), url(r'^enhancements/release/chart/?$', views.enhGraph, name='releasechart'), url(r'^enhancements/sprint/?$', views.SprintReport, name='sprint'), url(r'^enhancements/pending/?$', views.PendingUAT, name='UAT'), url(r'^enhancements/backlog/?$', views.Backlog, name='backlog'), url(r'^enhancements/backlog/(?P<chartType>[a-zA-Z]+)/?$', views.BacklogGraphs, name='blgraphs'), url(r'^enhancements/bymodule/?$', views.enhByModule, name='allbymod'), url(r'^enhancements/other/?$', views.NonFinance, name='nonfin'), url(r'^backlog/(?P<chartType>[a-zA-Z]+)/?$', views.BacklogGraphs, name='oldblgraphs'), url(r'^projects/grooming/?$', views.ProjectGrooming, name='projectGrooming'), url(r'^projects/active/(?P<epic>E[0-9]+)/?$', views.ProjectStories, name='projectStories'), url(r'^projects/active/?$', views.EpicView, name='activeProjects'), url(r'^velocity/old/?$', views.OldVelocityChart, name='oldvelocity'), url(r'^velocity/?$', views.VelocityChart, name='velocity'), url(r'^sprint/?$', views.FullSprint, name='full_sprint'), url(r'^syncstory/?$', views.updateStory, name='updateStory'), url(r'^info/?$', views.Info, name='info'), ]
UTF-8
Python
false
false
1,525
py
49
urls.py
21
0.668197
0.666885
0
25
60
100
msh5/pipfile-sort
7,816,840,513,810
9d995078121f8703b8757e4e8ee42b56a1074b46
fcc5337a191d4214655f8c34c84674270c153aa5
/pipfile_sort/__init__.py
d660901d32413abfb6582396a808173febfa3446
[ "MIT" ]
permissive
https://github.com/msh5/pipfile-sort
6dd6bab03a5e69f80bf241e0ed4792f418fe3272
c196cf282c20e2aaefdce8ea8959faf638311ef6
refs/heads/master
2023-02-18T07:41:47.518111
2021-09-02T13:07:07
2021-09-02T13:07:07
223,908,588
10
2
MIT
false
2023-02-08T01:50:09
2019-11-25T09:20:43
2022-11-15T13:13:27
2023-02-08T01:50:09
44
10
2
5
Python
false
false
from click import command from click import option from click import version_option from plette import Pipfile from plette.pipfiles import PackageCollection import sys APP_VERSION = '0.2.2' PIPFILE_FILENAME = './Pipfile' PIPFILE_ENCODING = 'utf-8' @command() @version_option(version=APP_VERSION) @option('--exit-code', is_flag=True, help= 'change to behavior of exit code. default behavior of return value, 0 is no differences, 1 is error exit. ' 'return 2 when add this option. 2 is exists differences.') def main(exit_code): # Load current data. with open(PIPFILE_FILENAME, encoding=PIPFILE_ENCODING) as f: pipfile = Pipfile.load(f) # Sort "dev-packages" mapping. sorted_dev_packages, all_changed = __sort_collection(pipfile.dev_packages) # Sort "packages" mapping. sorted_packages, changed = __sort_collection(pipfile.packages) if changed: all_changed = True # Replace with sorted lists pipfile.dev_packages = sorted_dev_packages pipfile.packages = sorted_packages # Store sorted data. with open(PIPFILE_FILENAME, 'w', encoding=PIPFILE_ENCODING) as f: Pipfile.dump(pipfile, f) # When --exit-code option is valid and package collection has been changed, exit with 2. if exit_code and all_changed: sys.exit(2) def __sort_collection(org_collection): org_packages = [p for p in org_collection] sorted_packages = sorted(org_packages) return ( PackageCollection({ p: org_collection[p]._data for p in sorted_packages }), org_packages != sorted_packages, )
UTF-8
Python
false
false
1,612
py
7
__init__.py
2
0.687345
0.681141
0
53
29.396226
111
vert-rouge/sample-zodb-migration-2to3
14,834,817,076,671
59339f1ab85776fd6bac63f3449b26ba9c844192
302c9d74cd6b5dd910d0d810d8f646e054d6662c
/readDbfrompython3.py
f85beb975702c42e212a1b27ff96932a7f740f2d
[]
no_license
https://github.com/vert-rouge/sample-zodb-migration-2to3
ff7f45cbdbcbac4d1f180a3abbf0424473c21d61
0295f7e9a5349135b18aad7dfd7b72a9a10237e1
refs/heads/master
2021-09-10T01:21:24.714389
2018-03-20T14:17:18
2018-03-20T14:17:18
126,027,959
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
# -*- coding: utf-8 -*- import ZODB, ZODB.FileStorage storage = ZODB.FileStorage.FileStorage('Data_python2.fs') db = ZODB.DB(storage) with db.transaction() as connection: print("la transaction est ouverte!") print(connection.root.truc) print(dir(connection.root.truc)) for k,v in connection.root.truc.items(): print(k) print(v)
UTF-8
Python
false
false
364
py
3
readDbfrompython3.py
3
0.67033
0.664835
0
15
23.266667
57
stevenman42/DefenderDotPyToo
9,680,856,298,397
05522959a044a6a662c665a5be72edb1642009c7
eb7e62574e80a3561d4917aef98d2e4cadd88dd7
/main.py
fe859ed44ee8094d2ca4fb07fb7a133e5471bb1c
[]
no_license
https://github.com/stevenman42/DefenderDotPyToo
6fbff160787c614c3f99350e9d24a8970608ccd4
e11bf1e1fba3509837b55f8ac74b738bc23760df
refs/heads/master
2020-05-07T17:42:24.301865
2014-07-10T02:01:46
2014-07-10T02:01:46
null
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
import pygame, sys from classes import * from math import * pygame.init() pygame.mixer.init() pygame.font.init() #test WIDTH = 960 HEIGHT = 640 BLUE = ( 50, 80,220) RED = (230, 70, 90) SKY_BLUE = (191,244,255) screen = pygame.display.set_mode((WIDTH,HEIGHT)) pygame.display.set_caption("Defender") clock = pygame.time.Clock() # pygame.event.set_grab(1) pygame.mouse.set_visible(0) EverythingList = [] # not necessarily everything, just everything that needs to be rendered! projectileList = [] CloudList = [] MortarList = [] TankList = [] EnemyList = [] HeartList = [] player = Player((960/2),(640-32),16,32) EverythingList.append(player) shootCount = 0 gravity = 2 totalFrames = 0 canShoot = True Score = 0 Lives = 10 Money = 0 MortarIcon = pygame.image.load("images/mortar_icon.png") LaserIcon = pygame.image.load("images/laser_icon.png") ground = pygame.image.load("images/ground.png") tank = pygame.image.load("images/tank.png") helicopter = pygame.image.load("images/helicopter.png") laser = pygame.mixer.Sound("sounds/shoot.wav") mortarSound = pygame.mixer.Sound("sounds/mortar.wav") MortarDrop = pygame.mixer.Sound("sounds/mortardrop.wav") HeliHit = pygame.mixer.Sound("sounds/helihit.wav") font = pygame.font.Font(None, 36) ActiveWeapon = 'laser' ActiveWeaponIcon = LaserIcon def GetMousePos(): mPos = pygame.mouse.get_pos() return mPos def GetMouseState(): clict = pygame.mouse.get_pressed()[0] return clict def Shoot(): global Money if ActiveWeapon == 'laser': laser.play(loops = 0) projectile = Projectile(player.xPos, player.yPos, 8, 8) if projectile.xPos < mPos[0]: projectile.xVel = (mPos[0]-player.xPos)/(uniform(10.5,11.5)) elif projectile.xPos > mPos[0]: projectile.xVel = (mPos[0]-player.xPos)/(uniform(10.5,11.5)) if projectile.yPos < mPos[1]: projectile.yVel = ((HEIGHT-(HEIGHT-player.yPos)) - mPos[1])/(uniform(10.5,11.5)) elif projectile.yPos > mPos[1]: projectile.yVel = -((HEIGHT-(HEIGHT-player.yPos)) - mPos[1])/(uniform(10.5,11.5)) projectileList.append(projectile) EverythingList.append(projectile) y = HEIGHT-(HEIGHT-player.yPos)-mPos[1] x = mPos[0]-player.xPos elif ActiveWeapon == 'mortar': if Money >= 5: mortarSound.play(loops = 0) mortProject = Projectile(player.xPos, player.yPos, 16, 16) if mortProject.xPos < mPos[0]: mortProject.xVel = (mPos[0]-player.xPos)/(uniform(9.5,10.5)) elif mortProject.xPos > mPos[0]: mortProject.xVel = (mPos[0]-player.xPos)/(uniform(9.5,10.5)) if mortProject.yPos < mPos[1]: mortProject.yVel = ((HEIGHT-(HEIGHT-player.yPos)) - mPos[1])/(uniform(9.5,10.5)) elif mortProject.yPos > mPos[1]: mortProject.yVel = -((HEIGHT-(HEIGHT-player.yPos)) - mPos[1])/(uniform(9.5,10.5)) MortarList.append(mortProject) EverythingList.append(mortProject) Money -= 5 elif ActiveWeapon == 'triple laser': laser.play(loops = 0) for i in [10,12,14]: projectile = Projectile(player.xPos, player.yPos, 8, 8) if projectile.xPos < mPos[0]: projectile.xVel = (mPos[0]-player.xPos)/i elif projectile.xPos > mPos[0]: projectile.xVel = (mPos[0]-player.xPos)/i if projectile.yPos < mPos[1]: projectile.yVel = ((HEIGHT-(HEIGHT-player.yPos)) - mPos[1])/i elif projectile.yPos > mPos[1]: projectile.yVel = -((HEIGHT-(HEIGHT-player.yPos)) - mPos[1])/i projectileList.append(projectile) EverythingList.append(projectile) y = HEIGHT-(HEIGHT-player.yPos)-mPos[1] x = mPos[0]-player.xPos while True: # Keyboard Input # for event in pygame.event.get(): if event.type == pygame.QUIT: pygame.quit() sys.exit() elif event.type == pygame.KEYDOWN: if event.key == pygame.K_LEFT: pass # Maybe player movement sometime in the future, but probably not elif event.key == pygame.K_RIGHT: pass elif event.key == pygame.K_ESCAPE: pygame.quit() sys.exit() elif event.key == pygame.K_1: ActiveWeapon = 'laser' ActiveWeaponIcon = LaserIcon elif event.key == pygame.K_2: ActiveWeapon = 'mortar' ActiveWeaponIcon = MortarIcon elif event.key == pygame.K_3: ActiveWeapon = 'triple laser' elif event.type == pygame.KEYUP: if event.key == pygame.K_LEFT: pass # End Keyboard Input # clict = GetMouseState() mPos = GetMousePos() # the stuff that happens when you click if clict == 1: # If the user left-clicks: mPos = GetMousePos() # Retrieve the position of the cursor for processing if canShoot == True: Shoot() canShoot = False else: pass if clict == 0: canShoot = True screen.blit(ground, (0, 0)) # Text (Score, Lives, Money!) # ScoreText = font.render(str(Score), 1, (10, 10, 10)) screen.blit(ScoreText, (10,10)) LivesText = font.render(str(Lives), 1, (255,10,10)) screen.blit(LivesText, (10, 40)) MoneyText = font.render(str(Money), 1, (80,205,50)) screen.blit(MoneyText, (920, 10)) # Weapon Icons # screen.blit(ActiveWeaponIcon, (10, 70)) for cloud in CloudList: if cloud.xPos >= WIDTH + 128: CloudList.remove(cloud) EverythingList.remove(cloud) for projectile in projectileList: if projectile.yPos < 0 or projectile.xPos < 0 or projectile.xPos > WIDTH: projectileList.remove(projectile) EverythingList.remove(projectile) print(len(projectileList)) for enemy in EnemyList: if enemy.xPos > -32 and enemy.xPos < WIDTH: enemy.rect = pygame.Rect(enemy.xPos,enemy.yPos,enemy.width,enemy.height) projectile.rect = pygame.Rect(projectile.xPos,projectile.yPos,projectile.width,projectile.height) if projectile.rect.colliderect(enemy.rect) or enemy.rect.contains(projectile.rect): if enemy in EnemyList: enemy.yVel = 18 HeliHit.play(loops = 0) if randint(1,15) == 1: NewHeart = Heart(enemy.xPos, enemy.yPos, 16,16) HeartList.append(NewHeart) EverythingList.append(NewHeart) if projectile in projectileList: projectileList.remove(projectile) EverythingList.remove(projectile) Score += 1 Money += 5 if enemy.xPos >= WIDTH + 64: EnemyList.remove(enemy) EverythingList.remove(enemy) if enemy.yPos >= HEIGHT: EnemyList.remove(enemy) EverythingList.remove(enemy) else: pass for heart in HeartList: heart.rect = pygame.Rect(heart.xPos,heart.yPos,heart.width,heart.height) projectile.rect = pygame.Rect(projectile.xPos,projectile.yPos,projectile.width,projectile.height) if projectile.rect.colliderect(heart.rect): if heart in HeartList: HeartList.remove(heart) EverythingList.remove(heart) Lives += 1 for mortar in MortarList: mortar.yVel += gravity if mortar.yPos > HEIGHT: MortarList.remove(mortar) EverythingList.remove(mortar) MortarDrop.play(loops = 0) for tank in TankList: tank.rect = pygame.Rect(tank.xPos,tank.yPos,tank.width,tank.height) mortar.rect = pygame.Rect(mortar.xPos,mortar.yPos,mortar.width,mortar.height) if mortar.rect.colliderect(tank.rect) or tank.rect.contains(mortar.rect): if tank in TankList: TankList.remove(tank) EverythingList.remove(tank) MortarList.remove(mortar) EverythingList.remove(mortar) Score += 2 Money += 10 for enemy in EnemyList: if enemy.xPos > WIDTH: EnemyList.remove(enemy) EverythingList.remove(enemy) Lives -= 1 for heart in HeartList: if heart.yPos > HEIGHT: HeartList.remove(heart) EverythingList.remove(heart) if totalFrames % 60 == 0: NCX = randint(-100, -50) NCY = randint(64, 128) NewCloud = Cloud(NCX, NCY, 128, 64) CloudList.append(NewCloud) EverythingList.append(NewCloud) if randint(1,2) == 1: NEX = randint(-100,0) NEY = randint(64,300) NewEnemy = Enemy(NEX, NEY, 32, 16,'helicopter') EnemyList.append(NewEnemy) EverythingList.append(NewEnemy) if randint(1,2) == 1: NEX = randint(-100,0) NEY = HEIGHT-16 NewEnemy = Enemy(NEX, NEY, 32, 16, 'tank') TankList.append(NewEnemy) EverythingList.append(NewEnemy) for tank in TankList: if tank.xPos > WIDTH/2: if tank in TankList: TankList.remove(tank) EverythingList.remove(tank) Lives -= 1 for everything in EverythingList: everything.render(screen) everything.xPos += everything.xVel everything.yPos += everything.yVel pygame.draw.rect(screen, (0,0,0), (mPos[0] - 16, mPos[1] - 1, 32, 2)) pygame.draw.rect(screen, (0,0,0), (mPos[0] - 1, mPos[1] - 16, 2, 32)) pygame.draw.circle(screen, (200, 60, 70), (mPos[0], mPos[1]), 2) pygame.draw.circle(screen, (0,0,0), (mPos[0], mPos[1]),12,2) clock.tick(60) shootCount += 1 totalFrames += 1 pygame.display.flip()
UTF-8
Python
false
false
8,608
py
4
main.py
3
0.682156
0.645795
0
340
24.320588
101
m-wrzr/code30
4,114,578,706,809
1152e8d9b88912fd1a70a0387411ab6311c98dff
1089f139ea13e422f8c52668866414cc1b18085d
/solutions/21/solution.py
2e97783123553a5579e1f6dd5341ad0761d2ee27
[]
no_license
https://github.com/m-wrzr/code30
d297d12888cf0256cf64daca594086f4cadb8938
2f28371685364b4b64e3253bd3a07d211f81ceee
refs/heads/master
2020-12-24T19:37:05.939264
2016-05-12T15:56:08
2016-05-12T15:56:08
56,145,017
1
0
null
null
null
null
null
null
null
null
null
null
null
null
null
inputSize = "large" with open("A-" + inputSize + "-practice.in", "r") as file: output = "" nCase = int(file.readline()) for case in range(0, nCase): inputs = file.readline().split() standing = 0 friendsNeeded = 0 for i, amount in enumerate([int(s) for s in list(inputs[1])]): friendsNeeded = max(friendsNeeded, i - standing) standing += amount output += "Case #{}: {}\n".format(case + 1, friendsNeeded) print(output) # write to txt with open("A-" + inputSize + "-output.txt", "w") as text_file: text_file.write(output)
UTF-8
Python
false
false
610
py
37
solution.py
18
0.570492
0.562295
0
24
24.291667
70
Zilby/Stuy-Stuff
10,307,921,511,886
db17878d614dc60e52483374b4d9f9052fec6ed5
2b25aae9266437b657e748f3d6fea4db9e9d7f15
/graphics/polygons/8/adam_dehovitz/draw.py
90b55664497a1feccbc227a858e47511f42846e6
[]
no_license
https://github.com/Zilby/Stuy-Stuff
b1c3bc23abf40092a8a7a80e406e7c412bd22ae0
5c5e375304952f62667d3b34b36f0056c1a8e753
refs/heads/master
2020-05-18T03:03:48.210196
2018-11-15T04:50:03
2018-11-15T04:50:03
24,191,397
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
from display import * from matrix import * import math def add_rect( points, x0, y0, z0, h, w, d): add_edge(points, x0, y0, z0, x0, y0, z0) add_edge(points, x0, y0 - h, z0, x0, y0 - h, z0) add_edge(points, x0, y0, z0 - d, x0, y0, z0 - d) add_edge(points, x0, y0 - h, z0 - d, x0, y0 - h, z0 - d) add_edge(points, x0 + w, y0 - h, z0, x0 + w, y0 - h, z0) add_edge(points, x0 + w, y0 - h, z0 - d,x0 + w, y0 - h, z0 - d) add_edge(points, x0 + w, y0, z0, x0 + w, y0, z0) add_edge(points, x0 + w, y0, z0 - d, x0 + w, y0, z0 - d) def add_sphere( points, cx, cy, r): step = 0.1 t1 = 0 while (t1 <= 1.00000001): t2 = 0 while (t2 <= 1.00000001): theta1 = math.pi * t1 theta2 = 2 * math.pi * t2 x = cx + r * math.cos(theta1) y = cy + r * math.sin(theta1) * math.cos(theta2) z = r * math.sin(theta1) * math.sin(theta2) add_point( points, x, y, z ) t2 += 0.01 t1 += 0.01 def add_torus(points, cx, cy, r1, r2): t1 = 0 while (t1 <= 1.001): t2 = 0 while (t2 <= 1.001): theta1 = 2 * math.pi * t1 theta2 = 2 * math.pi * t2 x = cx + math.cos(theta1) * (r1 * math.cos(theta2) + r2) y = cy + r1 * math.sin(theta2) z = math.sin(theta1) * (r1 * math.cos(theta2) + r2) add_point(points, x, y, z) t2 += 0.01 t1 += 0.01 def add_circle( points, cx, cy, cz, r, step ): x0 = cx + r y0 = cy z0 = cz t=0 while ( t<=1.0001) : x = r * math.cos( 2 * math.pi * t ) + cx y = r * math.sin( 2 * math.pi * t ) + cy z = cz #print( str(x) + ", " + str(y) ) add_edge (points, x0, y0, z0, x, y, z) x0 = x y0 = y z0 = z t= t+ step def add_curve( points, x0, y0, x1, y1, x2, y2, x3, y3, step, curve_type ): t = 0 xco = generate_curve_coefs(x0, x1, x2, x3, curve_type) yco = generate_curve_coefs(y0, y1, y2, y3, curve_type) while (t <= 1.0001): x = xco[0]*math.pow(t,3) + xco[1]*math.pow(t,2) + xco[2]*(t)+xco[3] y = yco[0]*math.pow(t,3) + yco[1]*math.pow(t,2) + yco[2]*(t)+yco[3] add_edge( points, x0, y0, 0, x, y, 0) x0 = x y0 = y t += step def draw_lines( matrix, screen, color ): if len( matrix ) < 2: print "Need at least 2 points to draw a line" p = 0 while p < len( matrix ) - 1: draw_line( screen, matrix[p][0], matrix[p][1], matrix[p+1][0], matrix[p+1][1], color ) p+= 2 def add_edge( matrix, x0, y0, z0, x1, y1, z1 ): add_point( matrix, x0, y0, z0 ) add_point( matrix, x1, y1, z1 ) def add_point( matrix, x, y, z=0 ): matrix.append( [x, y, z, 1] ) def draw_line( screen, x0, y0, x1, y1, color ): dx = x1 - x0 dy = y1 - y0 if dx + dy < 0: dx = 0 - dx dy = 0 - dy tmp = x0 x0 = x1 x1 = tmp tmp = y0 y0 = y1 y1 = tmp if dx == 0: y = y0 while y <= y1: plot(screen, color, x0, y) y = y + 1 elif dy == 0: x = x0 while x <= x1: plot(screen, color, x, y0) x = x + 1 elif dy < 0: d = 0 x = x0 y = y0 while x <= x1: plot(screen, color, x, y) if d > 0: y = y - 1 d = d - dx x = x + 1 d = d - dy elif dx < 0: d = 0 x = x0 y = y0 while y <= y1: plot(screen, color, x, y) if d > 0: x = x - 1 d = d - dy y = y + 1 d = d - dx elif dx > dy: d = 0 x = x0 y = y0 while x <= x1: plot(screen, color, x, y) if d > 0: y = y + 1 d = d - dx x = x + 1 d = d + dy else: d = 0 x = x0 y = y0 while y <= y1: plot(screen, color, x, y) if d > 0: x = x + 1 d = d - dy y = y + 1 d = d + dx
UTF-8
Python
false
false
4,306
py
742
draw.py
601
0.397817
0.33372
0
158
26.246835
75
HectorTR/ica
7,765,300,884,587
e86f03f7d907a6f68eb4d3818b8b81ff141a3ab4
37f0da68f87247c814fa6ad039ceee8be2167ffd
/my_test.py
d51f2e34d78d7b4119215e1fbe4356b9ce31deaf
[]
no_license
https://github.com/HectorTR/ica
7b4d65e31aec3bd7a0e09401b1763cd1e1e590d1
4caed51fb54175948edd07010d823b286b7f4236
refs/heads/master
2022-11-17T08:48:41.018118
2020-07-03T19:13:05
2020-07-03T19:13:05
262,254,893
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
import os def find(criteria) : finder = criteria.split("=") if len(finder) < 2 : raise Exception("find(criteria='key=value'): criteria syntax error") objects = gridlabd.get("objects") result = [] for name in objects : item = gridlabd.get_object(name) if finder[0] in item and item[finder[0]] == finder[1] : if "name" in item.keys() : result.append(item["name"]) else : result.append("%s:%s" % (item["class"],item["id"])) return result def on_init(t): houses = find("class=house") global recorder recorder = open("house.csv","w") recorder.write("name,datetime,temperature\n") return True def record_house(name,t) : global recorder if recorder : house = gridlabd.get_object(name) recorder.write("%s,%s,%s\n" % (house["name"],house["clock"],house["air_temperature"])) return True def on_term(t): recorder.close()
UTF-8
Python
false
false
955
py
4
my_test.py
3
0.591623
0.587435
0
33
27.909091
92
bopopescu/Stock_Predictor
17,575,006,207,099
e1dd81b061ab9a8e9a15ebf2b11005becd30898d
6f9b457083216409be606d9903586738504336ad
/env/lib/python2.7/sre_parse.py
c173415e47fe11ff527ea10fff4531f59bd548cb
[]
no_license
https://github.com/bopopescu/Stock_Predictor
ab998a3f83a0968cdc6a97a67298693932ae4537
bea27ceb06aeebb9d2f45ddf9dd17a6732a456c8
refs/heads/master
2022-04-12T10:42:34.313159
2018-12-09T05:04:09
2018-12-09T05:04:09
null
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
/Users/jeffreychen/anaconda2/lib/python2.7/sre_parse.py
UTF-8
Python
true
false
55
py
108
sre_parse.py
25
0.836364
0.781818
0
1
55
55
listar0810/jupyterlab-snippets-multimenus
17,712,445,159,228
71a220e15c2f5471e39ad1841d9df2f99dcc9bf7
9b32771b7d1513ee37bc62dd347675abcfc1bfc9
/example_snippets/multimenus_snippets/Snippets/NumPy/File IO/Write single array to text file.py
201b833875f14aa3b988fe7fb756335b515b2511
[ "BSD-3-Clause" ]
permissive
https://github.com/listar0810/jupyterlab-snippets-multimenus
44087ef1aeb030a3074862a337508b57d50072c6
477f51cfdbad7409eab45abe53cf774cd70f380c
refs/heads/master
2022-12-12T18:19:25.221083
2020-09-08T01:11:01
2020-09-08T01:11:01
null
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
np.savetxt(filename, x)
UTF-8
Python
false
false
23
py
884
Write single array to text file.py
881
0.782609
0.782609
0
1
23
23
muhadyan/Modul-2-JCDS07
15,315,853,398,982
7dc1ed5ddc1aaf48cfa39dc275cde1a08b9de9b1
93935b1f31257d80222d39113d54ffe06ba1c6fa
/Day_20/.ipynb_checkpoints/Day20Task-checkpoint.py
372f237dd99e9d99e3b686c39274394248d22fca
[]
no_license
https://github.com/muhadyan/Modul-2-JCDS07
5b34beea506d40a788898cfd4016eb81484b8c93
d6d32e3be6a36d20597f9ef9d04898990002ee01
refs/heads/master
2021-01-05T01:03:30.442952
2020-02-16T03:12:20
2020-02-16T03:12:20
240,823,002
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
import pandas as pd import numpy as np import matplotlib.pyplot as plt import mpl_toolkits.mplot3d.axes3d df = pd.read_excel('indo_12_1.xls', header=3, index_col=0, skipfooter=3, na_values='-') fig = plt.figure(figsize=[15,5]) p = plt.subplot(111, projection='3d') for j in df: for i in df: x = np.arange(len(df[i])) y = i z = np.zeros(len(df[i])) dx = np.ones(len(df[i])) dy = np.ones(len(df[i])) dz = df[i] p.bar3d(x, y, z, dx, dy, dz) plt.xticks(np.arange(len(df.index)), df.index, rotation=90) # plt.yticks(np.arange(len(list(df))), list(df), rotation=45) p.set_zlabel('Sumbu Z') plt.show()
UTF-8
Python
false
false
660
py
76
Day20Task-checkpoint.py
31
0.60303
0.572727
0
26
24.423077
87
urgemag/urge
4,483,945,892,610
0ab262cc8b6ec8c3cb8211990cfeea8d1223d785
a00f31c439f81fba6327a43f2cb0aeeace4cdea6
/flask/views/admin/accesses_add.py
cb94b0df5328f04bfb10f380f9b264897a5a78f6
[ "MIT" ]
permissive
https://github.com/urgemag/urge
183b6579262a2bf5f959ad193734e9da3683f0d7
a2ca0a9d2253770204dafbbad0cf9538c357d717
refs/heads/master
2023-04-21T15:44:45.721486
2021-04-26T20:15:47
2021-04-26T20:15:47
329,867,661
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
from flask import Blueprint, session, redirect, render_template, request, flash, url_for,abort from models import PageDetails, Database, General, Authentication from functools import wraps from validator_collection import * admin_add = Blueprint("admin_add", __name__) @admin_add.route("/Admin", methods=["POST", "GET"]) def admin_index(): return render_template("admin/admin_options.html") @admin_add.route("/Admin/Add", methods=["POST", "GET"]) def admin_add_index(): return render_template("admin/admin_add_options.html") @admin_add.route("/Admin/Social", methods=["POST", "GET"]) def admin_social_index(): return render_template("admin/admin_social_options.html") @admin_add.route("/Admin/Add/User", methods=["POST", "GET"]) def add_user_admin(): """ The Add User Page as an admin. """ message = False if request.method == "POST": accesses = [] for course in Database().get_all_slug_and_names_of_courses_from_db(): if (course["Slug"]) == request.form.get(course["Slug"]): accesses.append(course["Slug"]) message = Database().add_users_data_to_db( email=request.form.get("email"), password=General().sha256_hash(request.form.get("password")), first_name=request.form.get("name"), last_name=request.form.get("last_name"), phone_number=request.form.get("phone_num"), about_me=request.form.get("about_me"), bio=request.form.get("bio"), website=request.form.get("web"), birth_day=request.form.get("day"), birth_month=request.form.get("month"), birth_year=request.form.get("year"), Accesses=accesses, ) if message is True: message = {"Color": "green", "Result": "با موفقیت اضافه شد."} else: message["Result"] = message["Message"] message["Color"] = "red" flash(message) return redirect(url_for("admin_add.add_user_admin")) return render_template( "admin/admin_add_user.html", Accesses=PageDetails().all_accesses_admin_page(), message=message, ) @admin_add.route("/Admin/Add/Day", methods=["POST", "GET"]) def add_day_options_admin(): """ The Add options Page as an admin. """ return render_template( "admin/admin_add_day_options.html", ) @admin_add.route("/Admin/Add/Day/Essential", methods=["POST", "GET"]) def add_day_essential_main_data_admin(): """ The Add Essential Day Page as an admin. """ if request.method == "POST": def form_handler(request): slug = request.form.get("slug") day = request.form.get("day") name_day = request.form.get("name_day") description_status_existence = request.form.get("description-status") description = request.form.get("description") price_status = request.form.get("price-status") if is_not_empty(slug): return {"Result": False, "Message": "نام انگلیسی دوره را وارد کنید."} if is_not_empty(day) == "": return {"Result": False, "Message": "روز دوره را وارد کنید."} if is_not_empty(name_day): return {"Result": False, "Message": "اسم روز را وارد کنید."} if description_status_existence == "With-Description" and is_not_empty(description): return {"Result": False, "Message": "توضیحات روز را وارد کنید."} if price_status == "None-Free": freeness = False else: freeness = True if is_numeric(day): return {"Result": False, "Message": "روز دوره باید عدد باشد."} try: uploaded_file = request.files["cover"] except: return {"Result": False, "Message": "تصویر را آپلود کنید."} if Database().get_courses_data_from_db(request.form.get("slug")) is False: return {"Result": False, "Message": "همچین دوره ای وجود ندارد."} uploaded_image = request.files.get("cover") uploaded_image_bytes = uploaded_image.read() hash_image = General().sha256_hash_bytes(uploaded_image_bytes) format_file = General().format_recognizer(uploaded_image_bytes) file_name = "cover_of_day_of_course_" + str(hash_image) + "." + format_file if General().check_existence_of_a_file("static/assets/courses/{slug}/days/{day}".format(slug=slug,day=day)) is False: General().setup_course_folder(slug) location_image = "static/assets/courses/{slug}/days/{day}/{file_name}".format(slug=slug,day=day, file_name=file_name) location_image_href = "/static//assets/courses/{slug}/days/{day}/{file_name}".format(slug=slug,day=day, file_name=file_name) with open(location_image, "wb") as file: file.write(uploaded_image_bytes) General().image_resizer_using_imgp(location_image, 700) General().image_optimizer_using_imgp(location_image) message = Database().add_day_essential_data_to_db( course_name_slug=request.form.get("slug"), day_num=request.form.get("day"), name_persian=request.form.get("name_day"), description=request.form.get("description"), image_path=location_image_href, freeness=freeness, ) return message message = form_handler(request) if message is True: message = {"Color": "green", "Result": "با موفقیت اضافه شد."} else: if message["Result"] is False: message["Color"] = "red" else: message["Color"] = "green" message["Result"] = message["Message"] flash(message) return redirect(url_for("admin_add.add_day_essential_main_data_admin")) return render_template( "admin/admin_add_day_essential.html", Courses=Database().get_all_slug_and_names_of_courses_from_db(), ) import re @admin_add.route("/Admin/Add/Day/Text", methods=["POST", "GET"]) def add_day_text_data_admin(): """ The Add Main Text Day Page as an admin. """ if request.method == "POST": def form_handler(request): text = request.form.get("text") day = request.form.get("day") slug = request.form.get("slug") if slug == "": return {"Result": False, "Message": "نام انگلیسی دوره را وارد کنید."} if request.form.get("day") == "": return {"Result": False, "Message": "روز دوره را وارد کنید."} if text == "": return {"Result": False, "Message": "متن اصلی را وارد کنید."} try: int(day) except ValueError: return {"Result": False, "Message": "روز دوره باید عدد باشد."} if Database().get_courses_data_from_db(slug) is False: return {"Result": False, "Message": "همچین دوره ای وجود ندارد."} links_images = re.findall("src=[\"\'](.*?)[\"\']", text) if General().check_existence_of_a_file("static/assets/courses/{slug}/days/{day}".format(slug=slug,day=day)) is False: General().setup_course_folder(slug) for link in links_images: file_path = General().href_to_path(link) file_name = (file_path.split("/"))[-1] new_file_path = "static/assets/courses/{slug}/days/{day_number}/{file_name}".format(slug=slug,day_number=day, file_name=file_name) new_file_href = "/static//assets/courses/{slug}/days/{day_number}/{file_name}".format(slug=slug,day_number=day, file_name=file_name) General().move_file_path(file_path, new_file_path) text = text.replace(link, new_file_href) message = Database().add_day_text_data_to_db( course_name_slug=slug, day_num=day, text=text, ) return message message = form_handler(request) if message is True: message = {"Color": "green", "Result": "با موفقیت اضافه شد."} else: if message["Result"] is False: message["Color"] = "red" else: message["Color"] = "green" message["Result"] = message["Message"] flash(message) return redirect(url_for("admin_add.add_day_text_data_admin")) return render_template( "admin/admin_add_day_text.html", Courses=Database().get_all_slug_and_names_of_courses_from_db(), ) @admin_add.route("/Admin/Add/Day/Todo", methods=["POST", "GET"]) def add_day_todo_data_admin(): """ The Add Todo-Excersices Day Page as an admin. """ if request.method == "POST": def form_handler(request): if request.form.get("slug") == "": return {"Result": False, "Message": "نام انگلیسی دوره را وارد کنید."} if request.form.get("day") == "": return {"Result": False, "Message": "روز دوره را وارد کنید."} if request.form.get("todo") == "": return {"Result": False, "Message": "تمارین را وارد کنید."} try: int(request.form.get("day")) except ValueError: return {"Result": False, "Message": "روز دوره باید عدد باشد."} if Database().get_courses_data_from_db(request.form.get("slug")) is False: return {"Result": False, "Message": "همچین دوره ای وجود ندارد."} message = Database().add_day_todo_data_to_db( course_name_slug=request.form.get("slug"), day_num=request.form.get("day"), todo=request.form.get("todo"), ) return message message = form_handler(request) if message is True: message = {"Color": "green", "Result": "با موفقیت اضافه شد."} else: if message["Result"] is False: message["Color"] = "red" else: message["Color"] = "green" message["Result"] = message["Message"] flash(message) return redirect(url_for("admin_add.add_day_text_data_admin")) return render_template( "admin/admin_add_day_todo.html", Courses=Database().get_all_slug_and_names_of_courses_from_db(), ) @admin_add.route("/Admin/Add/Day/Quotes", methods=["POST", "GET"]) def add_day_quotes_data_admin(): """ The Add Quotes Day Page as an admin. """ if request.method == "POST": def form_handler(request): if request.form.get("slug") == "": return {"Result": False, "Message": "نام انگلیسی دوره را وارد کنید."} if request.form.get("day") == "": return {"Result": False, "Message": "روز دوره را وارد کنید."} if list(filter(None, request.form.getlist("quote"))) == []: return {"Result": False, "Message": "نقل قول ها را وارد کنید."} quotes = list(filter(None, request.form.getlist("quote"))) try: int(request.form.get("day")) except ValueError: return {"Result": False, "Message": "روز دوره باید عدد باشد."} if Database().get_courses_data_from_db(request.form.get("slug")) is False: return {"Result": False, "Message": "همچین دوره ای وجود ندارد."} message = Database().add_day_quotes_data_to_db( course_name_slug=request.form.get("slug"), day_num=request.form.get("day"), quotes=quotes, ) return message message = form_handler(request) if message is True: message = {"Color": "green", "Result": "با موفقیت اضافه شد."} else: if message["Result"] is False: message["Color"] = "red" else: message["Color"] = "green" message["Result"] = message["Message"] flash(message) return redirect(url_for("admin_add.add_day_quotes_data_admin")) return render_template( "admin/admin_add_day_quotes.html", Courses=Database().get_all_slug_and_names_of_courses_from_db(), ) @admin_add.route("/Admin/Add/Day/Music", methods=["POST", "GET"]) def add_day_music_data_admin(): """ The Add Music Day Page as an admin. """ if request.method == "POST": def form_handler(request): covers = request.files.getlist("cover") musics = request.files.getlist("music") creators = request.form.getlist("creator") names = request.form.getlist("music_name") if request.form.get("slug") == "": return {"Result": False, "Message": "نام انگلیسی دوره را وارد کنید."} if request.form.get("day") == "": return {"Result": False, "Message": "روز دوره را وارد کنید."} if request.form.get("description") == "": return {"Result": False, "Message": "توضیحات موزیک را وارد کنید."} for music in musics: if music.filename == "": musics.remove(music) for cover in covers: if cover.filename == "": covers.remove(cover) for name in names: if name == "": names.remove(name) for creator in creators: if creator == "": creators.remove(creator) if creators == [] or names == []: return {"Result": False, "Message": "اطلاعات موزیک را وارد کنید."} if musics == [] or covers == []: return { "Result": False, "Message": "مدیاهای مربوط به موزیک را وارد کنید.", } if not ( len(covers) == len(musics) and len(musics) == len(creators) and len(creators) == len(names) ): return {"Result": False, "Message": "همه فیلد های موزیک را وارد کنید."} try: int(request.form.get("day")) except ValueError: return {"Result": False, "Message": "روز دوره باید عدد باشد."} if Database().get_courses_data_from_db(request.form.get("slug")) is False: return {"Result": False, "Message": "همچین دوره ای وجود ندارد."} message = Database().add_day_musics_data_to_db( course_name_slug=request.form.get("slug"), day_num=request.form.get("day"), description=request.form.get("description"), covers=covers, musics=musics, creators=creators, names=names, ) return message message = form_handler(request) if message is True: message = {"Color": "green", "Result": "با موفقیت اضافه شد."} else: if message["Result"] is False: message["Color"] = "red" else: message["Color"] = "green" message["Result"] = message["Message"] flash(message) return redirect(url_for("admin_add.add_day_music_data_admin")) return render_template( "admin/admin_add_day_music.html", Courses=Database().get_all_slug_and_names_of_courses_from_db(), ) @admin_add.route("/Admin/Add/Day/Ted", methods=["POST", "GET"]) def add_day_ted_data_admin(): """ The Add TED video Day Page as an admin. """ if request.method == "POST": def form_handler(request): urls = request.form.getlist("ted_url") qualities = request.form.getlist("ted_quality") if request.form.get("slug") == "": return {"Result": False, "Message": "نام انگلیسی دوره را وارد کنید."} if request.form.get("day") == "": return {"Result": False, "Message": "روز دوره را وارد کنید."} if request.form.get("description") == "": return {"Result": False, "Message": "توضیحات ویدیوها را وارد کنید."} for url in urls: if url == "": urls.remove(url) for quality in qualities: if quality == "": qualities.remove(quality) if urls == [] or qualities == []: return {"Result": False, "Message": "لینک و یا کیفیت ها را وارد کنید."} if not (len(urls) == len(qualities)): return {"Result": False, "Message": "همه فیلد های ویدیو را وارد کنید."} try: int(request.form.get("day")) except ValueError: return {"Result": False, "Message": "روز دوره باید عدد باشد."} if Database().get_courses_data_from_db(request.form.get("slug")) is False: return {"Result": False, "Message": "همچین دوره ای وجود ندارد."} message = Database().add_day_ted_data_to_db( course_name_slug=request.form.get("slug"), day_num=request.form.get("day"), description=request.form.get("description"), urls=urls, qualities=qualities, ) return message message = form_handler(request) if message is True: message = {"Color": "green", "Result": "با موفقیت اضافه شد."} else: if message["Result"] is False: message["Color"] = "red" else: message["Color"] = "green" message["Result"] = message["Message"] flash(message) return redirect(url_for("admin_add.add_day_ted_data_admin")) return render_template( "admin/admin_add_day_ted.html", Courses=Database().get_all_slug_and_names_of_courses_from_db(), ) @admin_add.route("/Admin/Add/Day/Animation", methods=["POST", "GET"]) def add_day_animation_data_admin(): """ The Add short Animation film Day Page as an admin. """ if request.method == "POST": def form_handler(request): urls = request.form.getlist("animation_url") qualities = request.form.getlist("animation_quality") if request.form.get("slug") == "": return {"Result": False, "Message": "نام انگلیسی دوره را وارد کنید."} if request.form.get("day") == "": return {"Result": False, "Message": "روز دوره را وارد کنید."} if request.form.get("description") == "": return {"Result": False, "Message": "توضیحات ویدیو را وارد کنید."} for url in urls: if url == "": urls.remove(url) for quality in qualities: if quality == "": qualities.remove(quality) if urls == [] or qualities == []: return {"Result": False, "Message": "لینک و یا کیفیت ها را وارد کنید."} if not (len(urls) == len(qualities)): return { "Result": False, "Message": "همه فیلد های انیمیشن را وارد کنید.", } try: int(request.form.get("day")) except ValueError: return {"Result": False, "Message": "روز دوره باید عدد باشد."} if Database().get_courses_data_from_db(request.form.get("slug")) is False: return {"Result": False, "Message": "همچین دوره ای وجود ندارد."} message = Database().add_day_animation_data_to_db( course_name_slug=request.form.get("slug"), day_num=request.form.get("day"), description=request.form.get("description"), urls=urls, qualities=qualities, ) return message message = form_handler(request) if message is True: message = {"Color": "green", "Result": "با موفقیت اضافه شد."} else: if message["Result"] is False: message["Color"] = "red" else: message["Color"] = "green" message["Result"] = message["Message"] flash(message) return redirect(url_for("admin_add.add_day_animation_data_admin")) return render_template( "admin/admin_add_day_animation.html", Courses=Database().get_all_slug_and_names_of_courses_from_db(), ) @admin_add.route("/Admin/Add/Day/Podcast", methods=["POST", "GET"]) def add_day_podcast_data_admin(): """ The Add podcast Day Page as an admin. """ if request.method == "POST": def form_handler(request): creator = request.form.get("creator") name = request.form.get("podcast_name") cover = request.files.get("cover") url = request.form.get("podcast_url") if creator == "" or name == "": return {"Result": False, "Message": "اطلاعات پادکست را وارد کنید."} if request.form.get("slug") == "": return {"Result": False, "Message": "نام انگلیسی دوره را وارد کنید."} if request.form.get("day") == "": return {"Result": False, "Message": "روز دوره را وارد کنید."} if request.form.get("description") == "": return {"Result": False, "Message": "توضیحات پادکست را وارد کنید."} if url == "": return {"Result": False, "Message": "لینک پادکست را وارد کنید."} if cover.filename == "": return {"Result": False, "Message": "کاور پادکست را وارد کنید."} try: int(request.form.get("day")) except ValueError: return {"Result": False, "Message": "روز دوره باید عدد باشد."} if Database().get_courses_data_from_db(request.form.get("slug")) is False: return {"Result": False, "Message": "همچین دوره ای وجود ندارد."} message = Database().add_day_podcast_data_to_db( course_name_slug=request.form.get("slug"), day_num=request.form.get("day"), description=request.form.get("description"), url=url, cover=cover, creator=creator, name=name, ) return message message = form_handler(request) if message is True: message = {"Color": "green", "Result": "با موفقیت اضافه شد."} else: if message["Result"] is False: message["Color"] = "red" else: message["Color"] = "green" message["Result"] = message["Message"] flash(message) return redirect(url_for("admin_add.add_day_podcast_data_admin")) return render_template( "admin/admin_add_day_podcast.html", Courses=Database().get_all_slug_and_names_of_courses_from_db(), ) @admin_add.route("/Admin/Add/Day/Movie", methods=["POST", "GET"]) def add_day_movie_data_admin(): """ The Add Movie Day Page as an admin. """ if request.method == "POST": def form_handler(request): cover = request.files.get("cover") urls = request.form.getlist("movie_url") qualities = request.form.getlist("movie_quality") if request.form.get("slug") == "": return {"Result": False, "Message": "نام انگلیسی دوره را وارد کنید."} if request.form.get("day") == "": return {"Result": False, "Message": "روز دوره را وارد کنید."} if request.form.get("description") == "": return {"Result": False, "Message": "توضیحات پادکست را وارد کنید."} if cover.filename == "": return {"Result": False, "Message": "کاور پادکست را وارد کنید."} for url in urls: if url == "": urls.remove(url) for quality in qualities: if quality == "": qualities.remove(quality) if urls == [] or qualities == []: return {"Result": False, "Message": "لینک و یا کیفیت ها را وارد کنید."} if not (len(urls) == len(qualities)): return {"Result": False, "Message": "همه فیلد ها را وارد کنید."} try: int(request.form.get("day")) except ValueError: return {"Result": False, "Message": "روز دوره باید عدد باشد."} if Database().get_courses_data_from_db(request.form.get("slug")) is False: return {"Result": False, "Message": "همچین دوره ای وجود ندارد."} message = Database().add_day_movie_data_to_db( course_name_slug=request.form.get("slug"), day_num=request.form.get("day"), description=request.form.get("description"), urls=urls, qualities=qualities, cover=cover, ) return message message = form_handler(request) if message is True: message = {"Color": "green", "Result": "با موفقیت اضافه شد."} else: if message["Result"] is False: message["Color"] = "red" else: message["Color"] = "green" message["Result"] = message["Message"] flash(message) return redirect(url_for("admin_add.add_day_movie_data_admin")) return render_template( "admin/admin_add_day_movie.html", Courses=Database().get_all_slug_and_names_of_courses_from_db(), ) @admin_add.route("/Admin/Add/Course", methods=["POST", "GET"]) def add_course_admin(): """ The Add Course Page as an admin. """ if request.method == "POST": def form_course(request): if request.form.get("slug") == "": return {"Message": "نام انگلیسی دوره را وارد کنید."} slug = request.form.get("slug").replace(" ","-") try: uploaded_file = request.files["cover"] except: return {"Message": "تصویر را آپلود کنید."} try: length_of_course = int(request.form.get("len")) except ValueError: return {"Color": "red", "Message": "طول دوره باید عدد باشد."} if ( Database().get_courses_data_from_db(slug) != "" and uploaded_file.filename == "" ): return {"Message": "تصویر را آپلود کنید."} result_pic = General().save_picture_of_course( slug, uploaded_file, length_of_course ) if result_pic["Result"] is False: return result_pic if request.form.get("price-status") == "Free": price = "0" free = True else: price = request.form.get("price") free = False try: if request.form.get("soon") == "Soon": days_till_open = int(request.form.get("date_open")) else: days_till_open = 0 except ValueError: General().remove_file(result_pic["path"]) return {"Color": "red", "Message": "فاصله زمانی تا باز شدن باید عدد باشد."} try: int(price.replace(",", "")) except ValueError: General().remove_file(result_pic["path"]) return {"Color": "red", "Message": "قیمت دوره باید عدد باشد."} message = Database().add_course_data_to_db( name=request.form.get("name"), slug=slug, description=request.form.get("description"), intro=None, image_href=result_pic["href"], now_price=int(price.replace(",", "")), length_of_course=int(request.form.get("len")), robbin=request.form.get("robbin"), free=free, days_till_publish=days_till_open ) if message is not True: General().remove_file(result_pic["path"]) return message message = form_course(request) if message is True: message = {"Color": "green", "Result": "با موفقیت اضافه شد."} else: message["Result"] = message["Message"] message["Color"] = "red" flash(message) return redirect(url_for("admin_add.add_course_admin")) return render_template( "admin/admin_add_course.html", ) @admin_add.route("/Admin/Add/Course-Info", methods=["POST", "GET"]) def add_course_info_admin(): """ The Add Course information Page as an admin. """ if request.method == "POST": message = Database().add_course_info_to_db( slug=request.form.get("slug"), introduction=request.form.get("intro"), speciality=request.form.get("speciality"), importance=request.form.get("importance"), why=request.form.get("why"), length=request.form.get("length"), price=request.form.get("price"), last_words=request.form.get("last_word"), ) if message is True: message = {"Color": "green", "Result": "با موفقیت اضافه شد."} else: message["Result"] = message["Message"] message["Color"] = "red" flash(message) return redirect(url_for("admin_add.add_course_info_admin")) return render_template( "admin/admin_add_course_info.html", Courses=PageDetails().get_all_courses_info_categorized_by_info_existence(), ) @admin_add.route("/Admin/Add/Post", methods=["POST", "GET"]) def add_post_blog_admin(): """ The Add a Post for blog Page as an admin. """ if request.method == "POST": def form_handler(request): text = request.form.get("text") slug = (request.form.get("name_eng").replace(" ","-")).replace("_","-") if request.form.get("name_persian") == "": return {"Result": False, "Message": "نام فارسی دوره را وارد کنید."} if slug == "": return {"Result": False, "Message": "نام انگلیسی دوره را وارد کنید."} if text == "": return {"Result": False, "Message": "متن اصلی را وارد کنید."} uploaded_file = request.files["cover"] if uploaded_file.filename == "": return {"Message": "تصویر را آپلود کنید."} english_name = slug uploaded_image = request.files.get("cover") uploaded_image_bytes = uploaded_image.read() format_file = General().format_recognizer(uploaded_image_bytes) General().setup_blog_post_folder(slug) file_name = "blog-cover_" + english_name + "." + format_file location_image = "static/assets/images/blog/{}/".format(slug) + file_name location_image_href = "/static//assets/images/blog/{}/".format(slug) + file_name with open(location_image, "wb") as file: file.write(uploaded_image_bytes) General().image_resizer_using_imgp(location_image, 1500) General().image_optimizer_using_imgp(location_image) links_images = re.findall("src=[\"\'](.*?)[\"\']", text) for link in links_images: file_path = General().href_to_path(link) file_name = (file_path.split("/"))[-1] new_file_path = "static/assets/images/blog/{slug}/{file_name}".format(slug=slug, file_name=file_name) new_file_href = "/static//assets/images/blog/{slug}/{file_name}".format(slug=slug, file_name=file_name) General().move_file_path(file_path, new_file_path) text = text.replace(link, new_file_href) message = Database().add_post_blog_to_db( persian_name=request.form.get("name_persian"), eng_name=english_name, cover_href=location_image_href, text=text, ) return message message = form_handler(request) if message is True: message = {"Color": "green", "Result": "با موفقیت اضافه شد."} else: if message["Result"] is False: message["Color"] = "red" else: message["Color"] = "green" message["Result"] = message["Message"] flash(message) return redirect(url_for("admin_add.add_post_blog_admin")) return render_template( "admin/admin_add_post.html", ) @admin_add.route("/Admin/Add/Music", methods=["POST", "GET"]) def add_music_admin(): """ The Add Music as an admin. """ if request.method == "POST": def form_handler(request): cover = request.files.get("cover") music = request.files.get("music") creator = request.form.get("creator") name = request.form.get("music_name") if music.filename == "": return {"Result": False, "Message": "موزیک را اپلود کنید."} if cover.filename == "": return {"Result": False, "Message": "کاور را اپلود کنید."} if name == "" or creator == "": return {"Result": False, "Message": "اطلاعات موزیک را وارد کنید."} message = Database().add_music_data_to_db( cover=cover, music=music, creator=creator, name=name, ) return message message = form_handler(request) if message is True: message = {"Color": "green", "Result": "با موفقیت اضافه شد."} else: if message["Result"] is False: message["Color"] = "red" else: message["Color"] = "green" message["Result"] = message["Message"] flash(message) return redirect(url_for("admin_add.add_music_admin")) return render_template( "admin/admin_add_music.html", ) @admin_add.route("/Admin/Add/tool", methods=["POST", "GET"]) @admin_add.route("/Admin/Add/Tool", methods=["POST", "GET"]) def add_tool_admin(): """ The Add Tool Page as an admin. """ if request.method == "POST": def form_course(request): slug = request.form.get("slug").replace(" ","-") if slug == "": return {"Message": "نام انگلیسی را وارد کنید."} try: uploaded_file = request.files["cover"] except: return {"Message": "تصویر را آپلود کنید."} if ( Database().get_courses_data_from_db(slug) != "" and uploaded_file.filename == "" ): return {"Message": "تصویر را آپلود کنید."} result_pic = General().save_picture_of_tool(slug, uploaded_file) if result_pic["Result"] is False: return result_pic if request.form.get("price-status") == "Free": price = "0" else: price = request.form.get("price") try: if request.form.get("soon") == "Soon": days_till_open = int(request.form.get("date_open")) else: days_till_open = 0 except ValueError: General().remove_file(result_pic["path"]) return {"Color": "red", "Message": "فاصله زمانی تا باز شدن باید عدد باشد."} try: int(price.replace(",", "")) except ValueError: General().remove_file(result_pic["path"]) return {"Color": "red", "Message": "قیمت باید عدد باشد."} message = Database().add_tool_to_db( persian_name=request.form.get("name"), slug=slug, description=request.form.get("description"), image_href=result_pic["href"], now_price=int(price.replace(",", "")), robbin=request.form.get("robbin"), price=price, days_till_publish=days_till_open ) if message is not True: General().remove_file(result_pic["path"]) return message message = form_course(request) if message is True: message = {"Color": "green", "Result": "با موفقیت اضافه شد."} else: message["Result"] = message["Message"] message["Color"] = "red" flash(message) return redirect(url_for("admin_add.add_tool_admin")) return render_template( "admin/admin_add_tool.html", ) @admin_add.route("/Admin/Generate/Post", methods=["POST", "GET"]) def generate_post(): return render_template( "admin/admin_generate_instagram_post.html", templates=General().open_json_file("static/assets/instagram/posts.json") )
UTF-8
Python
false
false
39,273
py
78
accesses_add.py
19
0.525702
0.525195
0
1,000
36.468
148
L-sky/Master_Thesis
11,235,634,489,777
9d1ae6a18ee2d19298f41e50bde3e0c4327e3af8
11e58f5519b17b8a9bee0336c1e0f10999f637ba
/elastic_tensor/elastic_coef.py
3a5c45ba52f5d55640298bb8a5300aca2d7e87d5
[ "MIT" ]
permissive
https://github.com/L-sky/Master_Thesis
57e646225edffcc614ebc278324ef7616838c110
609e3b1c81dfb2e13d86df106d81e2e56d32488d
refs/heads/master
2022-12-25T05:13:39.040748
2020-10-10T21:39:09
2020-10-10T21:39:09
259,278,932
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
import sympy from sympy import sqrt from elastic_tensor.rgaunt import rgaunt_p as rgaunt from functools import partial common = rgaunt(0, 0, 0, 0, 0, 0) rgaunt_011 = partial(rgaunt, l1=0, l2=1, l3=1) rgaunt_211 = partial(rgaunt, l1=2, l2=1, l3=1) rgaunt_022 = partial(rgaunt, l1=0, l2=2, l3=2) rgaunt_222 = partial(rgaunt, l1=2, l2=2, l3=2) rgaunt_422 = partial(rgaunt, l1=4, l2=2, l3=2) def A(alpha, beta, gamma, delta): return rgaunt_011(0, alpha, beta) * rgaunt_011(0, gamma, delta) * common def B(alpha, beta, gamma, delta, m): B_value = rgaunt_022(0, m, m) * (rgaunt_011(0, alpha, beta) * rgaunt_211(m, gamma, delta) + rgaunt_011(0, gamma, delta) * rgaunt_211(m, alpha, beta)) return B_value def D(alpha, beta, gamma, delta): quintet = [-2, -1, 0, +1, +2] D_value = 0 for mp in quintet: D_value += rgaunt_211(mp, alpha, beta) * rgaunt_211(mp, gamma, delta) * rgaunt_022(0, mp, mp) return D_value def E(alpha, beta, gamma, delta, m): quintet = [-2, -1, 0, +1, +2] E_value = 0 for m1 in quintet: for m2 in quintet: E_value += rgaunt_211(m1, alpha, beta) * rgaunt_211(m2, gamma, delta) * rgaunt_222(m, m1, m2) return E_value def H(alpha, beta, gamma, delta, m): quintet = [-2, -1, 0, +1, +2] H_value = 0 for m1 in quintet: for m2 in quintet: H_value += rgaunt_211(m1, alpha, beta) * rgaunt_211(m2, gamma, delta) * rgaunt_422(m, m1, m2) return H_value def sph_cov_mat(): B_n2 = partial(B, m=-2) B_n1 = partial(B, m=-1) B_0 = partial(B, m=0) B_p1 = partial(B, m=+1) B_p2 = partial(B, m=+2) E_n2 = partial(E, m=-2) E_n1 = partial(E, m=-1) E_0 = partial(E, m=0) E_p1 = partial(E, m=+1) E_p2 = partial(E, m=+2) H_n4 = partial(H, m=-4) H_n3 = partial(H, m=-3) H_n2 = partial(H, m=-2) H_n1 = partial(H, m=-1) H_0 = partial(H, m=0) H_p1 = partial(H, m=+1) H_p2 = partial(H, m=+2) H_p3 = partial(H, m=+3) H_p4 = partial(H, m=+4) sph_func = [A, B_n2, B_n1, B_0, B_p1, B_p2, D, E_n2, E_n1, E_0, E_p1, E_p2, H_n4, H_n3, H_n2, H_n1, H_0, H_p1, H_p2, H_p3, H_p4] mat = sympy.zeros(81, 21) triplet = [-1, 0, +1] for idx in range(81): alpha_p = idx // 27 beta_p = (idx % 27) // 9 gamma_p = (idx % 9) // 3 delta_p = idx % 3 alpha = alpha_p - 1 beta = beta_p - 1 gamma = gamma_p - 1 delta = delta_p - 1 for j, sph in enumerate(sph_func): mat[idx, j] = sph(alpha, beta, gamma, delta) return mat def sph_cov_mat_sub(): B_n2 = partial(B, m=-2) B_n1 = partial(B, m=-1) B_0 = partial(B, m=0) B_p1 = partial(B, m=+1) B_p2 = partial(B, m=+2) E_n2 = partial(E, m=-2) E_n1 = partial(E, m=-1) E_0 = partial(E, m=0) E_p1 = partial(E, m=+1) E_p2 = partial(E, m=+2) H_n4 = partial(H, m=-4) H_n3 = partial(H, m=-3) H_n2 = partial(H, m=-2) H_n1 = partial(H, m=-1) H_0 = partial(H, m=0) H_p1 = partial(H, m=+1) H_p2 = partial(H, m=+2) H_p3 = partial(H, m=+3) H_p4 = partial(H, m=+4) sph_func = [A, B_n2, B_n1, B_0, B_p1, B_p2, D, E_n2, E_n1, E_0, E_p1, E_p2, H_n4, H_n3, H_n2, H_n1, H_0, H_p1, H_p2, H_p3, H_p4] mat = sympy.zeros(21, 21) quadruplets = [(-1, -1, -1, -1), (-1, -1, -1, 0), (-1, -1, -1, +1), (-1, -1, 0, 0), (-1, -1, 0, +1), (-1, -1, +1, +1), (-1, 0, -1, 0), (-1, 0, -1, +1), (-1, 0, 0, 0), (-1, 0, 0, +1), (-1, 0, +1, +1), (-1, +1, -1, +1), (-1, +1, 0, 0), (-1, +1, 0, +1), (-1, +1, +1, +1), (0, 0, 0, 0), (0, 0, 0, +1), (0, 0, +1, +1), (0, +1, 0, +1), (0, +1, +1, +1), (+1, +1, +1, +1)] for idx, quadruplet in enumerate(quadruplets): for j, sph in enumerate(sph_func): mat[idx, j] = sph(*quadruplet) return mat def reconstruction_coef(): T_scaled = sympy.Matrix([ [1, 0, 0, -2 * sqrt(5) / 5, 0, -2 * sqrt(15) / 5, 4 / 5, 0, 0, -4 * sqrt(5) / 35, 0, -4 * sqrt(15) / 35, 0, 0, 0, 0, 9 / 35, 0, 6 * sqrt(5) / 35, 0, 3 * sqrt(35) / 35], [0, 0, sqrt(15) / 5, 0, 0, 0, 0, 0, 2 * sqrt(15) / 35, 0, 0, 0, 0, -3 * sqrt(70) / 70, 0, -9 * sqrt(10) / 70, 0, 0, 0, 0, 0], [0, sqrt(15) / 5, 0, 0, 0, 0, 0, 2 * sqrt(15) / 35, 0, 0, 0, 0, -3 * sqrt(35) / 35, 0, -3 * sqrt(5) / 35, 0, 0, 0, 0, 0, 0], [1, 0, 0, sqrt(5) / 5, 0, -sqrt(15) / 5, -2 / 5, 0, 0, -4 * sqrt(5) / 35, 0, 4 * sqrt(15) / 35, 0, 0, 0, 0, -12 / 35, 0, -6 * sqrt(5) / 35, 0, 0], [0, 0, 0, 0, sqrt(15) / 5, 0, 0, 0, 0, 0, -4 * sqrt(15) / 35, 0, 0, 0, 0, 0, 0, -3 * sqrt(10) / 70, 0, -3 * sqrt(70) / 70, 0], [1, 0, 0, -2 * sqrt(5) / 5, 0, 0, -2 / 5, 0, 0, 8 * sqrt(5) / 35, 0, 0, 0, 0, 0, 0, 3 / 35, 0, 0, 0, -3 * sqrt(35) / 35], [0, 0, 0, 0, 0, 0, 3 / 5, 0, 0, 3 * sqrt(5) / 35, 0, -3 * sqrt(15) / 35, 0, 0, 0, 0, -12 / 35, 0, -6 * sqrt(5) / 35, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 3 * sqrt(15) / 35, 0, 0, 0, 0, 0, 0, -3 * sqrt(10) / 70, 0, -3 * sqrt(70) / 70, 0], [0, 0, sqrt(15) / 5, 0, 0, 0, 0, 0, 2 * sqrt(15) / 35, 0, 0, 0, 0, 0, 0, 6 * sqrt(10) / 35, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 3 * sqrt(15) / 35, 0, 0, 0, 0, 0, 0, 6 * sqrt(5) / 35, 0, 0, 0, 0, 0, 0], [0, 0, sqrt(15) / 5, 0, 0, 0, 0, 0, -4 * sqrt(15) / 35, 0, 0, 0, 0, 3 * sqrt(70) / 70, 0, -3 * sqrt(10) / 70, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 3 / 5, 0, 0, -6 * sqrt(5) / 35, 0, 0, 0, 0, 0, 0, 3 / 35, 0, 0, 0, -3 * sqrt(35) / 35], [0, sqrt(15) / 5, 0, 0, 0, 0, 0, -4 * sqrt(15) / 35, 0, 0, 0, 0, 0, 0, 6 * sqrt(5) / 35, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 3 * sqrt(15) / 35, 0, 0, 0, 0, 3 * sqrt(70) / 70, 0, -3 * sqrt(10) / 70, 0, 0, 0, 0, 0], [0, sqrt(15) / 5, 0, 0, 0, 0, 0, 2 * sqrt(15) / 35, 0, 0, 0, 0, 3 * sqrt(35) / 35, 0, -3 * sqrt(5) / 35, 0, 0, 0, 0, 0, 0], [1, 0, 0, 4 * sqrt(5) / 5, 0, 0, 4 / 5, 0, 0, 8 * sqrt(5) / 35, 0, 0, 0, 0, 0, 0, 24 / 35, 0, 0, 0, 0], [0, 0, 0, 0, sqrt(15) / 5, 0, 0, 0, 0, 0, 2 * sqrt(15) / 35, 0, 0, 0, 0, 0, 0, 6 * sqrt(10) / 35, 0, 0, 0], [1, 0, 0, sqrt(5) / 5, 0, sqrt(15) / 5, -2 / 5, 0, 0, -4 * sqrt(5) / 35, 0, -4 * sqrt(15) / 35, 0, 0, 0, 0, -12 / 35, 0, 6 * sqrt(5) / 35, 0, 0], [0, 0, 0, 0, 0, 0, 3 / 5, 0, 0, 3 * sqrt(5) / 35, 0, 3 * sqrt(15) / 35, 0, 0, 0, 0, -12 / 35, 0, 6 * sqrt(5) / 35, 0, 0], [0, 0, 0, 0, sqrt(15) / 5, 0, 0, 0, 0, 0, 2 * sqrt(15) / 35, 0, 0, 0, 0, 0, 0, -9 * sqrt(10) / 70, 0, 3 * sqrt(70) / 70, 0], [1, 0, 0, -2 * sqrt(5) / 5, 0, 2 * sqrt(15) / 5, 4 / 5, 0, 0, -4 * sqrt(5) / 35, 0, 4 * sqrt(15) / 35, 0, 0, 0, 0, 9 / 35, 0, -6 * sqrt(5) / 35, 0, 3 * sqrt(35) / 35] ]) return T_scaled def deconstruction_coef(): T_inv_scaled_transposed = sympy.Matrix([ [1 / 9, 0, 0, -sqrt(5) / 18, 0, -sqrt(15) / 18, 1 / 9, 0, 0, -sqrt(5) / 18, 0, -sqrt(15) / 18, 0, 0, 0, 0, 1 / 8, 0, sqrt(5) / 12, 0, sqrt(35) / 24], [0, 0, sqrt(15) / 9, 0, 0, 0, 0, 0, sqrt(15) / 9, 0, 0, 0, 0, -sqrt(70) / 12, 0, -sqrt(10) / 4, 0, 0, 0, 0, 0], [0, sqrt(15) / 9, 0, 0, 0, 0, 0, sqrt(15) / 9, 0, 0, 0, 0, -sqrt(35) / 6, 0, -sqrt(5) / 6, 0, 0, 0, 0, 0, 0], [2 / 9, 0, 0, sqrt(5) / 18, 0, -sqrt(15) / 18, -1 / 9, 0, 0, -sqrt(5) / 9, 0, sqrt(15) / 9, 0, 0, 0, 0, -1 / 3, 0, -sqrt(5) / 6, 0, 0], [0, 0, 0, 0, sqrt(15) / 9, 0, 0, 0, 0, 0, -2 * sqrt(15) / 9, 0, 0, 0, 0, 0, 0, -sqrt(10) / 12, 0, -sqrt(70) / 12, 0], [2 / 9, 0, 0, -sqrt(5) / 9, 0, 0, -1 / 9, 0, 0, 2 * sqrt(5) / 9, 0, 0, 0, 0, 0, 0, 1 / 12, 0, 0, 0, -sqrt(35) / 12], [0, 0, 0, 0, 0, 0, 1 / 3, 0, 0, sqrt(5) / 6, 0, -sqrt(15) / 6, 0, 0, 0, 0, -2 / 3, 0, -sqrt(5) / 3, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, sqrt(15) / 3, 0, 0, 0, 0, 0, 0, -sqrt(10) / 6, 0, -sqrt(70) / 6, 0], [0, 0, sqrt(15) / 9, 0, 0, 0, 0, 0, sqrt(15) / 9, 0, 0, 0, 0, 0, 0, sqrt(10) / 3, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, sqrt(15) / 3, 0, 0, 0, 0, 0, 0, 2 * sqrt(5) / 3, 0, 0, 0, 0, 0, 0], [0, 0, sqrt(15) / 9, 0, 0, 0, 0, 0, -2 * sqrt(15) / 9, 0, 0, 0, 0, sqrt(70) / 12, 0, -sqrt(10) / 12, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 1 / 3, 0, 0, -sqrt(5) / 3, 0, 0, 0, 0, 0, 0, 1 / 6, 0, 0, 0, -sqrt(35) / 6], [0, sqrt(15) / 9, 0, 0, 0, 0, 0, -2 * sqrt(15) / 9, 0, 0, 0, 0, 0, 0, sqrt(5) / 3, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, sqrt(15) / 3, 0, 0, 0, 0, sqrt(70) / 6, 0, -sqrt(10) / 6, 0, 0, 0, 0, 0], [0, sqrt(15) / 9, 0, 0, 0, 0, 0, sqrt(15) / 9, 0, 0, 0, 0, sqrt(35) / 6, 0, -sqrt(5) / 6, 0, 0, 0, 0, 0, 0], [1 / 9, 0, 0, sqrt(5) / 9, 0, 0, 1 / 9, 0, 0, sqrt(5) / 9, 0, 0, 0, 0, 0, 0, 1 / 3, 0, 0, 0, 0], [0, 0, 0, 0, sqrt(15) / 9, 0, 0, 0, 0, 0, sqrt(15) / 9, 0, 0, 0, 0, 0, 0, sqrt(10) / 3, 0, 0, 0], [2 / 9, 0, 0, sqrt(5) / 18, 0, sqrt(15) / 18, -1 / 9, 0, 0, -sqrt(5) / 9, 0, -sqrt(15) / 9, 0, 0, 0, 0, -1 / 3, 0, sqrt(5) / 6, 0, 0], [0, 0, 0, 0, 0, 0, 1 / 3, 0, 0, sqrt(5) / 6, 0, sqrt(15) / 6, 0, 0, 0, 0, -2 / 3, 0, sqrt(5) / 3, 0, 0], [0, 0, 0, 0, sqrt(15) / 9, 0, 0, 0, 0, 0, sqrt(15) / 9, 0, 0, 0, 0, 0, 0, -sqrt(10) / 4, 0, sqrt(70) / 12, 0], [1 / 9, 0, 0, -sqrt(5) / 18, 0, sqrt(15) / 18, 1 / 9, 0, 0, -sqrt(5) / 18, 0, sqrt(15) / 18, 0, 0, 0, 0, 1 / 8, 0, -sqrt(5) / 12, 0, sqrt(35) / 24] ]) return T_inv_scaled_transposed.T
UTF-8
Python
false
false
9,563
py
32
elastic_coef.py
23
0.419847
0.242183
0
200
46.815
176
futureneer/simple_ur
9,964,324,128,182
e34057714ee618b349b02b506e543ce2badf67db
e3cf50c77945d624bd22b5a2f8fcea8c1f786311
/simple_ur_driver/nodes/simple_urx_rt_example.py
a768e380b61e0191e84c3ef9e038ed9ee3722bb8
[ "BSD-2-Clause" ]
permissive
https://github.com/futureneer/simple_ur
9a8f85a18d907c6193ccae682c583a06f153996a
0004c6450c70f3ceb06e0e96f26dde8479012c28
refs/heads/master
2016-09-15T06:28:53.478621
2015-06-27T03:56:33
2015-06-27T03:56:33
22,484,196
2
0
null
null
null
null
null
null
null
null
null
null
null
null
null
#!/usr/bin/env python import rospy import numpy as np from copy import deepcopy import urx import logging import socket import struct if __name__ == "__main__": # rob = urx.Robot("192.168.1.155", logLevel=logging.INFO) print 'connected to robot' sock = socket.create_connection(('192.168.1.155', 30003), timeout=0.5) print 'created socket' # if rob: # realtime_monitor = rob.get_realtime_monitor() # print 'created realtime monitor connection' rospy.sleep(1) while not rospy.is_shutdown(): sock.send(struct.pack("!i", 2)) rospy.sleep(.1) print 'sending' # for n in range(500): # vels = [0.0, 0.0, 0.0, 0.0, -0.5, 0.0] # acc = .1 # vels.append(acc) # timeout = .01 # vels.append(timeout) # prog = "speedl([{},{},{},{},{},{}], a={}, t_min={})\n".format(*vels) # # prog = "textmsg({})\n".format(n) # if type(prog) != bytes: # prog = prog.encode() # # print 'sending command ['+str(prog)+']' # sock.send(prog) # # print realtime_monitor.get_all_data(wait=False)['tcp'][2] # # sock.recv(1044) # rospy.sleep(.007) # rospy.sleep(1) sock.close() # rob.cleanup()
UTF-8
Python
false
false
1,259
py
19
simple_urx_rt_example.py
12
0.544877
0.499603
0
44
27.636364
78
WillWagnerIV/Python-and-Data-Science
13,073,880,479,252
0790878db4832faff9b8cd5f5dca61ea5a6525ad
9ba84a4f815b217d7f62d63f0c340917f02a4a33
/recipes/hw6pr2.py
ba388fd04b8ea9f1719c84b2511bcd165852ce61
[]
no_license
https://github.com/WillWagnerIV/Python-and-Data-Science
e69bfd75ee512d429df687e4be938ca7aa80b46a
259f5c48065275948efe6f3c368c77b82fe4c963
refs/heads/master
2020-03-22T06:33:53.738606
2018-07-04T00:10:21
2018-07-04T00:10:21
null
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
# # starting examples for cs35, week1 "Web as Input" # import requests import string import json """ Examples to run for problem 1: Web scraping, the basic command (Thanks, Prof. Medero!) # # basic use of requests: # url = "https://www.cs.hmc.edu/~dodds/demo.html" # try it + source result = requests.get(url) text = result.text # provides the source as a large string... # # try it for another site... # # # let's try the Open Google Maps API -- also provides JSON-formatted data # See the webpage for the details and allowable use # # Try this one by hand - what are its parts? # http://maps.googleapis.com/maps/api/distancematrix/json?origins=%22Claremont,%20CA%22&destinations=%22Seattle,%20WA%22&mode=%22walking%22 # # Take a look at the result -- perhaps using this nice site for editing + display: # # A nice site for json display and editing: https://jsoneditoronline.org/ # # """ # ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ # # Problem 1 # # ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ # # example of calling the google distance API # def google_api(Sources, Dests, Mode): """ Inputs: Sources is a list of starting places Dests is a list of ending places This function uses Google's distances API to find the distances from all Sources to all Dests. It saves the result to distances.json Problem: right now it only works with the FIRST of each list! """ print("Start of google_api") url="http://maps.googleapis.com/maps/api/distancematrix/json" if len(Sources) < 1 or len(Dests) < 1: print("Sources and Dests need to be lists of >= 1 city each!") return start = Sources[0] end = Dests[0] my_mode= Mode # driving, walking, biking for z in range(1,len(Sources)): start += "|" start += Sources[z] for z in range(1,len(Dests)): end += "|" end += Dests[z] inputs={"origins":start,"destinations":end,"mode":my_mode} result = requests.get(url,params=inputs) data = result.json() # print("data is", data) # # save this json data to the file named distances.json # filename_to_save = "distances.json" f = open( filename_to_save, "w" ) # opens the file for writing string_data = json.dumps( data, indent=2 ) # this writes it to a string f.write(string_data) # then, writes that string to a file f.close() # and closes the file print("\nFile", filename_to_save, "written.") # no need to return anything, since we're better off reading it from file later... return # # example of handling json data via Python's json dictionary # def json_process(): """ This function reads the json data from "distances.json" It should build a formatted table of all pairwise distances. _You_ decide how to format that table (better than JSON!) """ filename_to_read = "distances.json" f = open( filename_to_read, "r" ) string_data = f.read() JD = json.loads( string_data ) # JD == "json dictionary" # print("The unformatted data in", filename_to_read, "is\n\n", JD, "\n") # print("Accessing some components:\n") row0 = JD['rows'][0] # print("row0 is", row0, "\n") cell0 = row0['elements'][0] # print("cell0 is", cell0, "\n") distance_as_string = cell0['distance']['text'] # print("distance_as_string is", distance_as_string, "\n") # we may want to continue operating on the whole json dictionary # so, we return it: return JD def pTable(): JD = json_process() s = 0 # print('\n\n') # while i < (len(row0['elements'])): while s < (len(Sources)): row = JD['rows'][s] oAdd = JD['origin_addresses'][s] print("Drive Time from " + oAdd + " to:") for d in range(len(Dests)): dAdd = JD['destination_addresses'][d] elems = row['elements'][d] distance_as_string = elems['distance']['text'] duration_as_string = elems['duration']['text'] print("- "+dAdd+ " is", distance_as_string,"in "+ duration_as_string) s += 1 print() return # # We're scripting! # if True: """ here's everything we'd like to run, when the file runs... """ # Sources Sources = ["Claremont,CA","Seattle,WA","Philadelphia,PA"] # Destinations Dests = ["Seattle,WA","Miami+FL","Boston+MA"] if True: # do we want to run the API call? google_api(Sources, Dests, "Driving") # get a new JSON file and save it! # either way, we want to process the JSON file: json_process() pTable() # Sources Sources = ["Kinshasa,Congo","Libreville,Gabon"] # Destinations Dests = ["Cairo,Egypt","Freetown, Ivory Coast"] if True: # do we want to run the API call? google_api(Sources, Dests, "Walking") # get a new JSON file and save it! # either way, we want to process the JSON file: json_process() pTable() # https://maps.googleapis.com/maps/api/distancematrix/json?origins=Vancouver+BC|Seattle&destinations=San+Francisco|Victoria+BC&mode=bicycling&language=fr-FR&key=YOUR_API_KEY
UTF-8
Python
false
false
5,330
py
38
hw6pr2.py
22
0.59531
0.58743
0
204
25.122549
173
anhlbt/VietnameseOCR
10,960,756,551,607
2f7d67023971af4d23ba6deb8b6ce4c5167e18b7
be591c275830d6c4a66e22e7337dfa65bd732406
/generate_dataset.py
649b421e47ce5bf614727015e794431215203a51
[ "Apache-2.0" ]
permissive
https://github.com/anhlbt/VietnameseOCR
e100162f0489b817c74e7c1be000a22dbbbad421
ccc2fbda772f72aeba4667769096beb51eeca9f2
refs/heads/master
2021-07-15T04:10:00.947369
2021-07-01T05:35:02
2021-07-01T05:35:02
189,510,030
0
0
Apache-2.0
true
2019-05-31T01:59:46
2019-05-31T01:59:46
2019-05-28T05:11:22
2018-08-30T03:01:40
110,223
0
0
0
null
false
false
#!/usr/bin/python # -*- coding: utf-8 -*- from PIL import Image, ImageDraw, ImageFont import matplotlib.pyplot as plt import pandas as pd import numpy as np import json import os import random import sys from config import * class DataGenerator: def __init__(self): self.i = 0 self.log = [] self.errors = [] self.data_folder = DATASET_DIR self.font_list = FONT_LIST self.data_set_csv = DATASET_FILE self.characters = [] self.dataset_size = 0 def rgb2gray(self, rgb): return np.dot(rgb[..., :3], [0.299, 0.587, 0.114]) def get_list_characters(self): if len(self.characters) != 0: return self.characters else: characters = [] with open(CHARACTERS_SET) as cf: for r in cf: if ',,' in r: c = ',' else: _, c = r.split(',') characters.append(c.replace('\n', '')) self.characters = characters return characters def create_text_image(self, text, font_ttf, idx_category, font_size): try: image = Image.new("RGB", (IMG_WIDTH, IMG_HEIGHT), (255, 255, 255)) draw = ImageDraw.Draw(image) font = ImageFont.truetype(font_ttf, font_size) w, h = draw.textsize(text, font=font) draw.text(((IMG_WIDTH - w) / 2, (IMG_HEIGHT - h) / 2), text, (0, 0, 0), font=font) if SAVE_TEXT_IMAGE_TO_DISK: image.save(self.data_folder + str(idx_category) + '/' + str(self.i) + '.jpg') self.log.append({'font': font_ttf, 'image': str(self.i) + '.jpg'}) self.i = self.i + 1 return image except Exception as e: self.errors.append({'font': font_ttf, 'errors': str(e)}) return None def generate_data_set(self, text, idx_category): images = [] with open(self.font_list, 'r') as fonts: for font in fonts: if '#' not in font: for font_size in range(FONT_SIZE_MIN, FONT_SIZE_MAX + 1): image = self.create_text_image(text, font.replace('\n', ''), idx_category, font_size) if image != None: self.dataset_size = self.dataset_size + 1 images.append(image) self.i = 0 return images def generate_dataset(self): characters = self.get_list_characters() for idx, ch in enumerate(characters): if SAVE_TEXT_IMAGE_TO_DISK and not os.path.exists(self.data_folder + str(idx)): os.makedirs(self.data_folder + str(idx)) c_images = self.generate_data_set(ch, idx) print('.', end='') for ic in c_images: image = np.asarray(ic) image = self.rgb2gray(image) image = image.reshape(1, IMG_WIDTH * IMG_HEIGHT) with open(DATASET_FILE, 'ab') as df: image = np.concatenate((image, np.array([[int(idx)]])), axis=1) np.savetxt(df, image, delimiter=",", fmt="%d") if __name__ == "__main__": print('Generating dataset...') generator = DataGenerator() generator.generate_dataset() print('Text Image Dataset is generated:', DATASET_FILE_PATH)
UTF-8
Python
false
false
3,424
py
10
generate_dataset.py
6
0.517231
0.505841
0
97
34.298969
109
ZIP97/Green
6,786,048,352,275
7827b254a037e1dd627886732320998c10fb4acd
78a7fbbb28ee8c7e47c2f4c3f815241efe0a44b7
/temp/网易财经-年报解析.py
3142483131783115bbf76cb70f8dfefc5b930c92
[]
no_license
https://github.com/ZIP97/Green
0287f8551f261f5735ceab2cc44142822b3874d2
27fc5277ab59a94ccf4a468331d75aec44cb7056
refs/heads/master
2020-04-18T22:24:43.915693
2019-02-10T15:26:27
2019-02-10T15:26:27
167,792,880
2
0
null
null
null
null
null
null
null
null
null
null
null
null
null
# coding: utf-8 import numpy as np import pandas as pd import requests, os from bs4 import BeautifulSoup from multiprocessing import Pool from Proxy import get_proxy, getWeb, gen_proxy proxies = get_proxy() def get_write(i): ''' 获取每年年报,并写入 txt 文件 :param i: 年度报告列表第几个年报 ''' text = getWeb(df.href.iloc[i], Return="text", proxies=proxies) text = text.replace("&ensp;", "") # text = text.replace("\r<br/>", "") txtFile = r"{}{}_{}.txt".format(txtPath, code, df["公告标题"].iloc[i]) with open(txtFile, "w") as f: f.write(text) return text txtPath = "./txt/" csvPath = "./年度报告/" code = "601899" corp = "紫金矿业" # 解析 网易财经 定期报告表格,获取年报链接 url = "http://quotes.money.163.com/f10/gsgg_{},dqbg.html".format(code) soupTable = getWeb(url, proxies=proxies, Return="soup") result = soupTable.find_all("div", class_="tabs_panel") res = result[0].get_text().split("\n") res = list(filter(None, res))[:-1] href = ["http://quotes.money.163.com"+item.get("href") for item in result[0].find_all("a")][:-3] npRes = np.array(res).reshape(-1, 3) df = pd.DataFrame(npRes[1:, :], columns=npRes[0, :]) df["href"] = href df["releaseYear"] = df["公布日期"].map(lambda x: x[:4]) index = df["公告类型"]=="年度报告" df = df[index] df.to_csv("{}{}_{}.csv".format(csvPath, code, corp), encoding="gbk", index=False) # 多进程获取网页数据并保存 pool = Pool() temp = pool.imap_unordered(get_write, range(len(df))) texts = [item for item in temp] pool.close()
UTF-8
Python
false
false
1,599
py
836
网易财经-年报解析.py
14
0.638947
0.621622
0
47
29.659574
96
galegosimpatico/playgrnd
18,262,200,954,502
68d42fc602b8eaddc21a764a4e8dff586a3fe384
96e78090a515a531c2020e44b09734e01e8cd29d
/contest/tuenti/2015/2/almost.py
7c6143fd523f1811df968aa350736d432635f9f2
[]
no_license
https://github.com/galegosimpatico/playgrnd
8684e5465b3d6250d567b9b3fd38d01ff0aad20a
e2efe4908fce39a93ed1df7ae3bc1e0dd3a745b6
refs/heads/master
2019-04-30T01:32:28.112503
2015-05-06T07:03:57
2015-05-06T07:03:57
12,020,126
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
#!/usr/bin/env python import argparse import sys USE_COMPREHENSION = False def check_prime(number, primes_set): number_is_prime = True for prime_number in primes_set: if number % prime_number == 0: number_is_prime = False break return number_is_prime def grow_almost_primes_set(primes_set, almost_primes_set, new_prime, n): for prime_number in primes_set: almost_prime_number = new_prime * prime_number if almost_prime_number <= n: almost_primes_set.add(almost_prime_number) def create_almost_primes_set(n): if not USE_COMPREHENSION: # Init lists. primes_set = {2, 3} almost_primes_set = {4, 6, 9} # Generate all primes up to (n / 2). for i in range(5, int(n / 2) + 1): i_is_prime = check_prime(i, primes_set) if i_is_prime: primes_set.add(i) grow_almost_primes_set( primes_set, almost_primes_set, i, n) return sorted(almost_primes_set) else: primes_set = {x for x in range(2, n) if all( x % y for y in range(2, int(n**.5)))} sorted_primes_set = sorted(primes_set) almost_primes_set = { {x * y for x in sorted_primes_set} for y in sorted_primes_set} # almost_primes_generator = ( # (x * y for x in sorted_primes_set) for y in sorted_primes_set) # almost_primes_set = set(almost_primes_generator) # almost_primes_list = [ # [x * y for x in sorted_primes_set] for y in sorted_primes_set] # almost_primes_set = set(almost_primes_list) return sorted(almost_primes_set) def find_almost_primes_in_interval(almost_primes_set, min_, max_): almost_primes_in_interval = set() for almost_prime_number in almost_primes_set: if min_ <= almost_prime_number <= max_: almost_primes_in_interval.add(almost_prime_number) elif almost_prime_number > max_: break return almost_primes_in_interval def run(): parser = argparse.ArgumentParser() parser.add_argument('-i', '--input_file') parser.add_argument('-o', '--output_file') arguments = parser.parse_args() if arguments.input_file: input_file = open(arguments.input_file) cases_no = int(input_file.readline()) lines = input_file.readlines() else: cases_no = int(sys.stdin.readline()) lines = sys.stdin.readlines() if arguments.output_file: output_file = open(arguments.output_file, 'w') else: output_file = None # Find most bigger 'b' present in any test case. max_b = 6 for line in lines: candidate_b = int(line.split(' ')[1]) if candidate_b > max_b: max_b = candidate_b almost_primes_set = create_almost_primes_set(max_b) for case in range(cases_no): a, b = map(int, lines[case].split(' ')) almost_primes_in_interval = find_almost_primes_in_interval( almost_primes_set, a, b) case_result = almost_primes_in_interval.__len__() if arguments.output_file: output_file.write(str(case_result) + '\n') output_file.flush() else: print(case_result) sys.stdout.flush() if __name__ == '__main__': run()
UTF-8
Python
false
false
3,362
py
66
almost.py
33
0.580309
0.575848
0
105
31.019048
76
longshuo1995/craw_conf
1,511,828,500,742
5bc68c3d9d5a87ff2f4df62c13a0db5765891bbe
0dd0c04e495984b6ae46131fe360c597b1447915
/crawl_conf/flask_interface/02_interface_test.py
ad392753de8b2f973794522b4ffeba9f7859704b
[]
no_license
https://github.com/longshuo1995/craw_conf
9a20f7fdec1b821f0f1e7170619e7362a061fa0c
ae98e8fcf209bca8e959054e09621fba79864b02
refs/heads/master
2020-03-08T04:22:43.585495
2018-04-12T06:59:35
2018-04-12T06:59:35
127,919,371
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
import requests import json data = { "url": "http://www.baidu.com", "seed": { "article": { "field_name": "article", "field_rule": "//a/href", "rule_type": "xpath", "field_type": "article" }, "create_time": { "field_name": "create_time", "field_rule": "//span", "rule_type": "xpath", "field_type": "create_time" } } } data = {'data': json.dumps(data)} res = requests.post("http://127.0.0.1:10086/parse", data=data).text print(json.loads(res))
UTF-8
Python
false
false
615
py
29
02_interface_test.py
25
0.447154
0.429268
0
24
24.625
67
rogeriosilva-ifpi/adsi-algoritmos-2016.1
17,875,653,901,033
43ca87ce2dae9ce6449afed61985483aa86cf8ce
9861218f60ab23d6ac3bc6b400c220abf4e64fb5
/atividade_d/gerson/GERSON CAVALCANTE PROVA/q08.py
e760cb6f525802539359c55f4f03a77ebdb6e126
[]
no_license
https://github.com/rogeriosilva-ifpi/adsi-algoritmos-2016.1
a0b0709eb783110a9b335c8364aa41ce4f90fb24
1714e2480b80e46be4d96049e878bf17b692320b
refs/heads/master
2021-06-06T18:25:00.836715
2016-09-07T02:02:30
2016-09-07T02:02:30
null
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
a = input() b = input() c = input() if (b - c) < a < (b + c) and (a - c) < b < (a + c) and (a - b) < c < (a + b): perimetro = a + b + c print "Perimetro = %.1f " %perimetro else: area = ((a + b) * c )/2 print"Area : %.1f " %area
UTF-8
Python
false
false
240
py
284
q08.py
281
0.4375
0.425
0
9
24.888889
77
dfrojas/insumos
584,115,583,415
5ab13b14dd9e3b2d63bda3316a794cb4fe8037ba
80b73d863f25cdfbeb75ef0aff7ffe84266d868c
/agroecoturismo/models.py
3f209f988790222481aac683f0d1069102d3d3e7
[]
no_license
https://github.com/dfrojas/insumos
18354f37fce32230aaff7ff346e0b0068a9ebb21
1ecafe8393a40ff4dab1689235515f49a55d5e62
refs/heads/master
2016-09-09T20:48:04.899543
2015-04-28T21:01:27
2015-04-28T21:01:27
34,008,214
0
0
null
null
null
null
null
null
null
null
null
null
null
null
null
from django.db import models from predio.models import InfoPredioGeneral from opciones.models import AtractivoEcoturismo, ActividadEcoturismo class Agroecoturismo(models.Model): predio = models.ForeignKey(InfoPredioGeneral) atractivo = models.ManyToManyField(AtractivoEcoturismo) actividad = models.ManyToManyField(ActividadEcoturismo) observaciones = models.TextField() def __str__(self): return self.predio.nombre_predio
UTF-8
Python
false
false
433
py
74
models.py
67
0.82679
0.82679
0
13
32.307692
68
archerckk/PyTest
19,404,662,244,320
7b20318b7310321c53c80f2f6ea7a4db2fe42c9f
595c69f717fc3ceb4e0701cc433f6d7f927b6fdb
/Ar_Script/ar_045_字典_通讯录程序.py
323628e994e1a70f7d4c4d958cdf9b048add94cf
[ "MIT" ]
permissive
https://github.com/archerckk/PyTest
d6462ebf46c6dbd5bb3ce03666aad0c2665367cd
610dd89df8d70c096f4670ca11ed2f0ca3196ca5
refs/heads/master
2022-03-26T21:09:25.891745
2021-06-14T01:39:36
2021-06-14T01:39:36
129,497,345
0
0
null
false
2020-01-14T10:57:49
2018-04-14T08:23:03
2020-01-13T14:29:04
2020-01-14T10:57:49
28,113
0
0
0
Python
false
false
print() '尝试利用字典的特性编写一个通讯录程序' ''' 1.查询联系人资料 存在打印,不存在提示 2.插入新的联系人 不存在直接新增 已存在提示是否修改 是,覆盖原来的电话,提示修改成功 否,跳出当前新增 3.删除联系人 联系人不存在 提示删除的联系人不存在 联系人存在 提示是否删除 是,删除,打印删除id 否,跳出当前选项 4.退出通讯录程序 打印提示语,终止循环 ''' print( ''' |---欢迎进入通讯录程序---| |---1.查询联系人资料 ---| |---2.插入新的联系人 ---| |---3.删除已有联系人 ---| |---4.退出通讯录程序 ---|''' ) def contacts(): codeList=['1','2','3','4'] contact={} while 1: code=input('\n请输入相关的指令代码:') if code not in codeList: print('非法指令代码!',end='') if code=='1': name=input('请输入联系人姓名:') # if name not in contact: # print('你所查找的联系人不存在!!') # else: # print(name+':',contact[name]) '异常处理优化代码效率' try: print(name + ':', contact[name]) except KeyError: print('你所查找的联系人不存在!!') if code=='2': name = input('请输入联系人姓名:') if name not in contact: contact[name]=input('请输入联系方式:') else: print('你输入的姓名在通讯录中已存在-->>%s:%s'%(name,contact[name])) '两个以上的格式化参数一定要记得加()不然会说格式化参数的数量不对' while 1: decide = input('是否修改用户资料(YES/NO:)') if decide in ['YES','yes','Yes']: contact[name] = input('请输入联系方式:') print('修改资料成功') break if decide in ['NO','No''no']: break else: print('你的输入有误!',end='') if code=='3': name = input('请输入联系人姓名:') if name in contact: while 1: decide = input('是否删除用户资料(YES/NO:)') if decide in ['YES','yes','Yes']: del contact[name] print('删除用户%s:'%name) break if decide in ['NO','No','no']: break else: print('你的输入有误!',end='') else: print('你要删除的用户不存在!') if code=='4': print('|---感谢使用通讯录程序---|') break contacts()
UTF-8
Python
false
false
3,067
py
575
ar_045_字典_通讯录程序.py
530
0.397633
0.389303
0
91
24.076923
69
aferriss/dcganMod
17,360,257,844,229
366b67baa57efdba6ab9f8f5c33a4db68d243c1b
8a1ea67914acd3e6aa32005cd955da21d9f32f41
/test.py
15d6f9eb82be418da1b5d9d479b97fca5f25c226
[ "BSD-3-Clause" ]
permissive
https://github.com/aferriss/dcganMod
442e8bb76a05307f9c539326bbb03b4ebacdb408
68e0323b908c8aa7088350bd16c562a87b09e75d
refs/heads/master
2021-01-10T14:56:09.691317
2017-05-16T01:26:42
2017-05-16T01:26:42
53,920,361
1
1
null
false
2017-05-20T17:23:28
2016-03-15T06:26:42
2016-12-21T20:51:53
2017-05-16T01:26:57
3,036
0
1
1
Lua
null
null
import math import numpy as np import theano ny = 11 def floatX(X): return np.asarray(X, dtype=theano.config.floatX) def OneHot(X, n=None, negative_class=0.): print "neg class" print n X = np.asarray(X).flatten() print (X) if n is None: n = np.max(X) + 1 Xoh = np.ones((len(X), n)) * negative_class Xoh[np.arange(len(X)), X] = 1. return Xoh def get_buffer_y(steps,num_buffer_samples=10,num_buffer_steps=2): num_buffer_rows = int(math.ceil(float(num_buffer_samples) / steps)) print("num buf rows= " + str(num_buffer_rows)) targets = np.asarray([[int(round(i*steps+_/num_buffer_steps)) for _ in range(steps)] for i in range(num_buffer_rows)]) end = np.empty([num_buffer_rows, steps], dtype='int') for _ in range( num_buffer_rows): empty = np.empty(steps, dtype='int') for i in range(steps): end[_][i] = (int(round(_*steps+i/num_buffer_steps))) ymb = floatX(OneHot(end.flatten(), ny)) return ymb y = get_buffer_y(5) print y
UTF-8
Python
false
false
967
py
3
test.py
1
0.654602
0.645295
0
42
22.02381
119