I am trying to scrape LinkedIn website using Selenium. I can't parse Next button. It resists as much as it can. I've spent a half of a day to adress this, but all in vain.
I tried absolutely various options, with text and so on. Only work with start ID but scrape other button.
selenium.common.exceptions.NoSuchElementException: Message: no such element: Unable to locate element: {"method":"xpath","selector":"//button[@aria-label='Далее']"}
This is quite common for this site:
//*[starts-with(@id,'e')]
My code:
from selenium import webdriver from selenium.webdriver import Keys from selenium.webdriver.chrome.service import Service from selenium.webdriver.common.by import By from time import sleep chrome_driver_path = Service("E:\programming\chromedriver_win32\chromedriver.exe") driver = webdriver.Chrome(service=chrome_driver_path) url = "https://www.linkedin.com/feed/" driver.get(url) SEARCH_QUERY = "python developer" LOGIN = "EMAIL" PASSWORD = "PASSWORD" sleep(10) sign_in_link = driver.find_element(By.XPATH, '/html/body/div[1]/main/p[1]/a') sign_in_link.click() login_input = driver.find_element(By.XPATH, '//*[@id="username"]') login_input.send_keys(LOGIN) sleep(1) password_input = driver.find_element(By.XPATH, '//*[@id="password"]') password_input.send_keys(PASSWORD) sleep(1) enter_button = driver.find_element(By.XPATH, '//*[@id="organic-div"]/form/div[3]/button') enter_button.click() sleep(25) lens_button = driver.find_element(By.XPATH, '//*[@id="global-nav-search"]/div/button') lens_button.click() sleep(5) search_input = driver.find_element(By.XPATH, '//*[@id="global-nav-typeahead"]/input') search_input.send_keys(SEARCH_QUERY) search_input.send_keys(Keys.ENTER) sleep(5) people_button = driver.find_element(By.XPATH, '//*[@id="search-reusables__filters-bar"]/ul/li[1]/button') people_button.click() sleep(5) page_button = driver.find_element(By.XPATH, "//button[@aria-label='Далее']") page_button.click() sleep(60)