Skip to content
Merged
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
84 changes: 69 additions & 15 deletions lessons/02_web_scraping.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -47,19 +47,62 @@
},
{
"cell_type": "code",
"execution_count": null,
"execution_count": 1,
"metadata": {},
"outputs": [],
"source": [
"%pip install requests"
"outputs": [
{
"name": "stdout",
"output_type": "stream",
"text": [
"Requirement already satisfied: requests in c:\\users\\jjala\\appdata\\local\\programs\\python\\python313\\lib\\site-packages (2.32.5)\n",
"Requirement already satisfied: charset_normalizer<4,>=2 in c:\\users\\jjala\\appdata\\local\\programs\\python\\python313\\lib\\site-packages (from requests) (3.4.3)\n",
"Requirement already satisfied: idna<4,>=2.5 in c:\\users\\jjala\\appdata\\local\\programs\\python\\python313\\lib\\site-packages (from requests) (3.10)\n",
"Requirement already satisfied: urllib3<3,>=1.21.1 in c:\\users\\jjala\\appdata\\local\\programs\\python\\python313\\lib\\site-packages (from requests) (2.5.0)\n",
"Requirement already satisfied: certifi>=2017.4.17 in c:\\users\\jjala\\appdata\\local\\programs\\python\\python313\\lib\\site-packages (from requests) (2025.8.3)\n",
"Note: you may need to restart the kernel to use updated packages.\n"
]
}
],
"source": [
"# 🌐 La librería requests es necesaria para hacer solicitudes HTTP y descargar páginas web.\n",
"# 🕸️ Esto es fundamental para hacer web scraping (extraer información de páginas web).\n",
"%pip install requests "
]
},
{
"cell_type": "code",
"execution_count": null,
"metadata": {},
"outputs": [],
"source": [
"outputs": [
{
"name": "stdout",
"output_type": "stream",
"text": [
"Collecting beautifulsoup4\n",
" Downloading beautifulsoup4-4.13.4-py3-none-any.whl.metadata (3.8 kB)\n",
"Collecting soupsieve>1.2 (from beautifulsoup4)\n",
" Downloading soupsieve-2.7-py3-none-any.whl.metadata (4.6 kB)\n",
"Collecting typing-extensions>=4.0.0 (from beautifulsoup4)\n",
" Downloading typing_extensions-4.14.1-py3-none-any.whl.metadata (3.0 kB)\n",
"Downloading beautifulsoup4-4.13.4-py3-none-any.whl (187 kB)\n",
"Downloading soupsieve-2.7-py3-none-any.whl (36 kB)\n",
"Downloading typing_extensions-4.14.1-py3-none-any.whl (43 kB)\n",
"Installing collected packages: typing-extensions, soupsieve, beautifulsoup4\n",
"\n",
" ------------- -------------------------- 1/3 [soupsieve]\n",
" -------------------------- ------------- 2/3 [beautifulsoup4]\n",
" -------------------------- ------------- 2/3 [beautifulsoup4]\n",
" -------------------------- ------------- 2/3 [beautifulsoup4]\n",
" ---------------------------------------- 3/3 [beautifulsoup4]\n",
"\n",
"Successfully installed beautifulsoup4-4.13.4 soupsieve-2.7 typing-extensions-4.14.1\n",
"Note: you may need to restart the kernel to use updated packages.\n"
]
}
],
"source": [
"# 🥣 La instrucción %pip install beautifulsoup4 sirve para instalar la librería Beautiful Soup 4 en tu entorno de Jupyter Notebook.\n",
"# 🕸️ Beautiful Soup es esencial para analizar y extraer información de archivos HTML y XML, lo que facilita el web scraping.\n",
"%pip install beautifulsoup4"
]
},
Expand All @@ -72,9 +115,25 @@
},
{
"cell_type": "code",
"execution_count": null,
"execution_count": 3,
"metadata": {},
"outputs": [],
"outputs": [
{
"name": "stdout",
"output_type": "stream",
"text": [
"Collecting lxml\n",
" Downloading lxml-6.0.1-cp313-cp313-win_amd64.whl.metadata (3.9 kB)\n",
"Downloading lxml-6.0.1-cp313-cp313-win_amd64.whl (4.0 MB)\n",
" ---------------------------------------- 0.0/4.0 MB ? eta -:--:--\n",
" ----- ---------------------------------- 0.5/4.0 MB 5.7 MB/s eta 0:00:01\n",
" ---------------------------------------- 4.0/4.0 MB 15.9 MB/s 0:00:00\n",
"Installing collected packages: lxml\n",
"Successfully installed lxml-6.0.1\n",
"Note: you may need to restart the kernel to use updated packages.\n"
]
}
],
"source": [
"%pip install lxml"
]
Expand Down Expand Up @@ -988,7 +1047,7 @@
"metadata": {
"anaconda-cloud": {},
"kernelspec": {
"display_name": "Python 3 (ipykernel)",
"display_name": "Python 3",
"language": "python",
"name": "python3"
},
Expand All @@ -1002,12 +1061,7 @@
"name": "python",
"nbconvert_exporter": "python",
"pygments_lexer": "ipython3",
"version": "3.8.13"
},
"vscode": {
"interpreter": {
"hash": "b6f9fe9f4b7182690503d8ecc2bae97b0ee3ebf54e877167ae4d28c119a56988"
}
"version": "3.13.6"
}
},
"nbformat": 4,
Expand Down