Home
last modified time | relevance | path

Searched refs:soup (Results 1 – 23 of 23) sorted by relevance

/openbmc/openbmc/poky/bitbake/lib/bs4/tests/
H A Dtest_tree.py235 soup = self.soup(data)
685 soup = self.soup("")
713 soup = self.soup("")
719 soup = self.soup("")
861 soup = self.soup(
932 soup = self.soup("")
940 soup = self.soup("")
948 soup = self.soup(
975 soup = self.soup(
1040 soup = self.soup(
[all …]
H A Dtest_html5lib.py32 soup = self.soup(markup, parse_only=strainer)
34 soup.decode(), self.document_for(markup))
70 soup = self.soup(markup)
72 self.assertEqual(b"<p>foo</p>", soup.p.encode())
76 soup = self.soup(markup)
78 self.assertEqual(2, len(soup.find_all('p')))
83 soup = self.soup(markup)
85 self.assertEqual(2, len(soup.find_all('p')))
90 soup = self.soup(markup)
95 soup = self.soup(markup)
[all …]
H A Dtest_soup.py41 soup = self.soup(data)
46 soup = self.soup(data)
51 soup = self.soup(utf8_data, exclude_encodings=["utf-8"])
63 soup = self.soup("<a><b></b></a>")
69 soup = self.soup("<a><b></b></a>", "html")
75 soup = self.soup("<a><b></b></a>", "html.parser")
89 soup = self.soup(utf8, fromEncoding="utf8")
106 soup = self.soup(filename)
114 soup = self.soup(filename)
119 soup = self.soup("http://www.crummy.com/")
[all …]
H A Dtest_lxml.py50 soup = self.soup("<!DOCTYPE>")
51 doctype = soup.contents[0]
58 soup = BeautifulStoneSoup("<b />")
59 self.assertEqual("<b/>", str(soup.b))
H A Dtest_htmlparser.py26 tree = self.soup("<a><b>foo</a>")
/openbmc/openbmc/poky/bitbake/lib/bs4/
H A Dtesting.py96 soup = self.soup(markup)
133 soup = self.soup(markup)
140 soup = self.soup(markup)
157 soup = self.soup("<p/>")
198 soup = self.soup(html)
206 soup = self.soup(markup)
264 soup = self.soup(markup)
272 soup = self.soup(markup)
389 soup = self.soup(text)
429 soup = self.soup(text)
[all …]
H A Ddiagnose.py70 soup = BeautifulSoup(data, parser)
77 print(soup.prettify())
179 soup = BeautifulSoup(data, parser)
H A D__init__.py170 self.builder.soup = self
223 self.builder.soup = None
467 soup = BeautifulSoup(sys.stdin) variable
468 print(soup.prettify())
H A DNEWS.txt457 new <p> tag will look like "<p />" if the soup object was created to
458 parse XML, but it will look like "<p></p>" if the soup object was
689 constructed soup object should always be valid as part of an XML
714 arguments would replace it, but they don't. You can't call soup('a',
768 soup(args={"id" : "5"}) with soup(id="5"). You can still use args if
784 string or a filehandle into the soup constructor, not with feed after
785 the soup has been created. There is still a feed method, but it's the
959 soup.bodyTag.pTag.tableTag('th'):
967 This lets you do tree traversals like soup.htmlTag.headTag.titleTag
/openbmc/openbmc/poky/bitbake/lib/bs4/builder/
H A D_html5lib.py79 self.soup = soup
83 self.soup.reset()
84 return Element(self.soup, self.soup, None)
102 self.soup = BeautifulSoup("")
104 return Element(self.soup, self.soup, None)
108 self.soup.append(node.element)
111 return self.soup
150 self.soup = soup
234 text = TextNode(self.soup.new_string(data), self.soup)
332 def __init__(self, element, soup): argument
[all …]
H A D_htmlparser.py68 self.soup.handle_data(data)
97 self.soup.endData()
98 self.soup.handle_data(data)
99 self.soup.endData(Comment)
102 self.soup.endData()
108 self.soup.handle_data(data)
109 self.soup.endData(Doctype)
117 self.soup.endData()
119 self.soup.endData(cls)
122 self.soup.endData()
[all …]
H A D_lxml.py69 self.soup = None
186 self.soup.endData()
187 completed_tag = self.soup.tagStack[-1]
202 self.soup.endData()
207 self.soup.handle_data(content)
210 self.soup.endData()
212 self.soup.object_was_parsed(doctype)
216 self.soup.endData()
217 self.soup.handle_data(content)
218 self.soup.endData(Comment)
[all …]
H A D__init__.py99 self.soup = None
190 self.soup.handle_starttag(name, attrs)
194 self.soup.handle_endtag(name)
215 self.soup.handle_data(content)
/openbmc/openbmc/poky/meta/recipes-support/libsoup/libsoup-2.4/
H A D0001-Fix-build-with-libxml2-2.12.0-and-clang-17.patch14 libsoup/soup-xmlrpc-old.c | 1 +
15 libsoup/soup-xmlrpc.c | 1 +
18 diff --git a/libsoup/soup-xmlrpc-old.c b/libsoup/soup-xmlrpc-old.c
20 --- a/libsoup/soup-xmlrpc-old.c
21 +++ b/libsoup/soup-xmlrpc-old.c
29 #include "soup-xmlrpc-old.h"
30 diff --git a/libsoup/soup-xmlrpc.c b/libsoup/soup-xmlrpc.c
32 --- a/libsoup/soup-xmlrpc.c
33 +++ b/libsoup/soup-xmlrpc.c
40 #include "soup-xmlrpc.h"
[all …]
/openbmc/openbmc/meta-openembedded/meta-oe/recipes-extended/flatpak/
H A Dflatpak_1.15.6.bb55 PACKAGECONFIG[soup] = "-Dhttp_backend=soup,,libsoup-2.4"
/openbmc/openbmc/poky/meta/recipes-support/libsoup/
H A Dlibsoup_3.4.4.bb35 cat >${WORKDIR}/soup.cross <<EOF
40 EXTRA_OEMESON:append:class-target = " --cross-file ${WORKDIR}/soup.cross"
H A Dlibsoup-2.4_2.74.3.bb39 cat >${WORKDIR}/soup.cross <<EOF
44 EXTRA_OEMESON:append:class-target = " --cross-file ${WORKDIR}/soup.cross"
/openbmc/openbmc/poky/bitbake/lib/bb/fetch2/
H A Dwget.py478soup = BeautifulSoup(self._fetch_index(url, ud, d), "html.parser", parse_only=SoupStrainer("a"))
479 if not soup:
483 for line in soup.find_all('a', href=True):
527soup = BeautifulSoup(self._fetch_index(dirs_uri, ud, d), "html.parser", parse_only=SoupStrainer("a…
528 if not soup:
531 for line in soup.find_all('a', href=True):
/openbmc/openbmc/meta-openembedded/meta-oe/recipes-extended/ostree/
H A Dostree_2024.5.bb43 # soup for curl to avoid bringing in deprecated libsoup2 (though
54 # soup or curl - https://github.com/ostreedev/ostree/issues/1897
88 PACKAGECONFIG[soup2] = "--with-soup, --without-soup, libsoup-2.4, , , soup3"
/openbmc/openbmc/poky/meta/recipes-multimedia/gstreamer/
H A Dgstreamer1.0-plugins-good_1.22.11.bb21 RPROVIDES:${PN}-soup += "${PN}-souphttpsrc"
22 RDEPENDS:${PN}-soup += "${MLPREFIX}${@bb.utils.contains('PACKAGECONFIG', 'soup2', 'libsoup-2.4', 'l…
H A Dgstreamer1.0-meta-base.bb39 gstreamer1.0-plugins-good-soup"
/openbmc/openbmc/meta-openembedded/meta-networking/recipes-support/strongswan/
H A Dstrongswan_5.9.14.bb43 PACKAGECONFIG[soup] = "--enable-soup,--disable-soup,libsoup-2.4,${PN}-plugin-soup"
/openbmc/openbmc/poky/meta/lib/oe/
H A Ddistro_check.py19 soup = BeautifulSoup(create_socket(url,d), "html.parser", parse_only=SoupStrainer("a"))
21 for line in soup.find_all('a', href=True):