Great video!! Is there a way to extract the information of only the experience? I tried with your code but when extracting experience or education information the ' ' spaces are 18 for both topics.
Hello, I have a problem when extracting the data Warning (from warnings module): File "C:\Users\fredd\OneDrive\Imágenes\linkedin-web-scraping-main (1)\linkedin-web-scraping-main\linkedin_employee_scraper.py", line 11 driver = webdriver.Chrome(path) DeprecationWarning: executable_path has been deprecated, please pass in a Service object Warning (from warnings module): File "C:\Users\fredd\OneDrive\Imágenes\linkedin-web-scraping-main (1)\linkedin-web-scraping-main\linkedin_employee_scraper.py", line 42 source = BeautifulSoup(driver.page_source) GuessedAtParserWarning: No parser was explicitly specified, so I'm using the best available HTML parser for this system ("lxml"). This usually isn't a problem, but if you run this code on another system, or in a different virtual environment, it may use a different parser and behave differently. The code that caused this warning is on line 42 of the file C:\Users\fredd\OneDrive\Imágenes\linkedin-web-scraping-main (1)\linkedin-web-scraping-main\linkedin_employee_scraper.py. To get rid of this warning, pass the additional argument 'features="lxml"' to the BeautifulSoup constructor. Traceback (most recent call last): File "C:\Users\fredd\OneDrive\Imágenes\linkedin-web-scraping-main (1)\linkedin-web-scraping-main\linkedin_employee_scraper.py", line 176, in searchable = getProfileURLs(company) File "C:\Users\fredd\OneDrive\Imágenes\linkedin-web-scraping-main (1)\linkedin-web-scraping-main\linkedin_employee_scraper.py", line 53, in getProfileURLs title = invisibleguy.findNext('div', class_='lt-line-clamp lt-line-clamp--multi-line ember-view').contents[0].strip(' ').strip(' ') AttributeError: 'NoneType' object has no attribute 'contents' Greetings from Ecuador
this is a common problem; the program can't find the chromedriver path, so download the chromedriver exe that matches whichever chrome version you're using here: chromedriver.chromium.org/downloads then when downloaded you can just put the exe file in the same directory as the program and set the path string to be the name of the chromedriver exe
what's annoying is linkedin changes their ui frequently to prevent web scrapers like us from harvesting data. A program like this needs to be constantly maintained
That intro 😭 i felt ur pain thanks top G for such a precious help
Great video!! Is there a way to extract the information of only the experience? I tried with your code but when extracting experience or education information the ' ' spaces are 18 for both topics.
Can you also specify the format in which we need to write our login.txt in ?
Excellent video!👍
Do you know how to by pass the security verification check? It's giving me the verification prompt after several times running the program
Great work.
Github link please. Just now please
github.com/thelazyaz/linkedin-web-scraping
Hello, I have a problem when extracting the data
Warning (from warnings module):
File "C:\Users\fredd\OneDrive\Imágenes\linkedin-web-scraping-main (1)\linkedin-web-scraping-main\linkedin_employee_scraper.py", line 11
driver = webdriver.Chrome(path)
DeprecationWarning: executable_path has been deprecated, please pass in a Service object
Warning (from warnings module):
File "C:\Users\fredd\OneDrive\Imágenes\linkedin-web-scraping-main (1)\linkedin-web-scraping-main\linkedin_employee_scraper.py", line 42
source = BeautifulSoup(driver.page_source)
GuessedAtParserWarning: No parser was explicitly specified, so I'm using the best available HTML parser for this system ("lxml"). This usually isn't a problem, but if you run this code on another system, or in a different virtual environment, it may use a different parser and behave differently.
The code that caused this warning is on line 42 of the file C:\Users\fredd\OneDrive\Imágenes\linkedin-web-scraping-main (1)\linkedin-web-scraping-main\linkedin_employee_scraper.py. To get rid of this warning, pass the additional argument 'features="lxml"' to the BeautifulSoup constructor.
Traceback (most recent call last):
File "C:\Users\fredd\OneDrive\Imágenes\linkedin-web-scraping-main (1)\linkedin-web-scraping-main\linkedin_employee_scraper.py", line 176, in
searchable = getProfileURLs(company)
File "C:\Users\fredd\OneDrive\Imágenes\linkedin-web-scraping-main (1)\linkedin-web-scraping-main\linkedin_employee_scraper.py", line 53, in getProfileURLs
title = invisibleguy.findNext('div', class_='lt-line-clamp lt-line-clamp--multi-line ember-view').contents[0].strip('
').strip(' ')
AttributeError: 'NoneType' object has no attribute 'contents'
Greetings from Ecuador
this is a common problem; the program can't find the chromedriver path, so download the chromedriver exe that matches whichever chrome version you're using here: chromedriver.chromium.org/downloads then when downloaded you can just put the exe file in the same directory as the program and set the path string to be the name of the chromedriver exe
Thanks ! Do not work anymore in June 2023
what's annoying is linkedin changes their ui frequently to prevent web scrapers like us from harvesting data. A program like this needs to be constantly maintained
@@aiwithaz yeah really hard ^^