Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 5 years ago.
Improve this question
I have a pdf file containing text, images and tables.I want to extract just the tables from that pdf file using either Python or R.
If you are considering using R I would recommend using the tabulizer package.
it is available here and is very easy to use.
to install it you would have to use the following command:
install.packages("devtools")
devtools::install_github("ropensci/tabulizer")
And using one of their examples:
library("tabulizer")
f <- system.file("examples", "data.pdf", package = "tabulizer")
# When f is your selected pdf file.
out1 <- extract_tables(f)
# Or even better, say what page the tables are in.
out2 <- extract_tables(f, pages = 1, guess = FALSE, method = "data.frame")
You'll probably find PyPI useful - you can search for specific things on there like 'PDF' and it will give you a list of modules relating to PDF's (here). You'll probably want PDF 1.0 judging from it's weight on PyPI. This should help you get started!
Related
Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 1 year ago.
Improve this question
I would like to ask you for a starting point or an idea to search for information on the subject. I have found many post about "how to run R script in Python", but I just need to use a function inside a specific package. So, I would like to know if it is possible to invoke a R function, using my data in Python and get the output. Thanks in advance!
I would look into rpy2. It takes a minute to set up but you can call R packages and use their functions fairly easily. All you have to do is use importr followed by the package you want to use from there. A few places to get started are:
https://rpy2.github.io/doc/v3.0.x/html/robjects_functions.html
https://medium.com/analytics-vidhya/calling-r-from-python-magic-of-rpy2-d8cbbf991571
Best of luck.
Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 2 years ago.
Improve this question
I have a sample image of a graph that is extracted from an excel file using a win32 package of python.
The implementation of the above is solved in this question = extracting graphs as images from excel
Now, I want to read an image in python and write it to a doc file and save it. which ultimately gives me an output of somethig like this
You can use the add_picture()-method of the following library:
python-docx
This also gives you options for resizing and placement.
Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 3 years ago.
Improve this question
There must be a simple solution to this question:
"Create a .pdf file from various other .pdf files w/ navigable index and page numbers via python."
All files are in the same folder, and all are .pdf files.
I want each filename to contain in the index and the index as a starting page.
What packages do you think fits my needs best?
Any thoughts?
https://pythonhosted.org/PyPDF2/
You have to split each page that you want and to add to your new pdf file.
Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 4 years ago.
Improve this question
One of the most popular tools to generate static sites is Sphinx which is largely used in the Python community to document code. It converts .rst files into other formats like HTML, PDF and others. But how is it possible that a static documentation with plain HTML files is searchable without losing performance?
I guess, it's done by creating an index (like a JSON file for example) that will be loaded via AJAX and is interpreted by something like lunr.js. Since many major projects in the world of Python have a huge documentation (like the Python docs itself). Therefore, how is it possible, to create such a good search without creating a gigantic index file that needs to be loaded?
You can use Google Search Engine to use Google´s power on your site. It is difficult to customize yet powerful. Other reference in this question
Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 8 years ago.
Improve this question
I've been searching for hours on how to extract the main text of a Wikipedia article, without all the links and references. I've tried wikitools, mwlib, BeautifulSoup and more. But I haven't really managed to.
Is there any easy and fast way for me to take the clear text (the actual article), and put it in a Python variable?
SOLUTION: Omid Raha solved it :)
You can use this package, that is a python wrapper for Wikipedia API,
Here is a quick start.
First install it:
pip install wikipedia
Example:
import wikipedia
p = wikipedia.page("Python programming language")
print(p.url)
print(p.title)
content = p.content # Content of page.
Output:
http://en.wikipedia.org/wiki/Python_(programming_language)
Python (programming language)